var/home/core/zuul-output/0000755000175000017500000000000015111516212014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111527205015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004566301215111527177017713 0ustar rootrootNov 26 06:25:15 crc systemd[1]: Starting Kubernetes Kubelet... Nov 26 06:25:15 crc restorecon[4673]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:15 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 06:25:16 crc restorecon[4673]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 26 06:25:17 crc kubenswrapper[4775]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 06:25:17 crc kubenswrapper[4775]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 26 06:25:17 crc kubenswrapper[4775]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 06:25:17 crc kubenswrapper[4775]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 06:25:17 crc kubenswrapper[4775]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 26 06:25:17 crc kubenswrapper[4775]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.059808 4775 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068755 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068797 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068806 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068814 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068823 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068831 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068839 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068847 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068855 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068863 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068871 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068879 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068886 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068894 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068902 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068910 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068918 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068926 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068937 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068948 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068956 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068965 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068974 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068982 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068991 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.068999 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069007 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069017 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069026 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069035 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069043 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069051 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069058 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069066 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069073 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069081 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069090 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069100 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069110 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069119 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069127 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069135 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069144 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069152 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069160 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069167 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069175 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069183 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069192 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069199 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069207 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069215 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069226 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069236 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069246 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069256 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069264 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069274 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069282 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069291 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069299 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069307 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069314 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069322 4775 feature_gate.go:330] unrecognized feature gate: Example Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069330 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069337 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069345 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069352 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069360 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069367 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.069375 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071227 4775 flags.go:64] FLAG: --address="0.0.0.0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071254 4775 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071274 4775 flags.go:64] FLAG: --anonymous-auth="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071286 4775 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071297 4775 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071306 4775 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071318 4775 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071331 4775 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071341 4775 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071351 4775 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071362 4775 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071373 4775 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071383 4775 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071392 4775 flags.go:64] FLAG: --cgroup-root="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071401 4775 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071410 4775 flags.go:64] FLAG: --client-ca-file="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071419 4775 flags.go:64] FLAG: --cloud-config="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071428 4775 flags.go:64] FLAG: --cloud-provider="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071437 4775 flags.go:64] FLAG: --cluster-dns="[]" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071447 4775 flags.go:64] FLAG: --cluster-domain="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071455 4775 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071465 4775 flags.go:64] FLAG: --config-dir="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071473 4775 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071483 4775 flags.go:64] FLAG: --container-log-max-files="5" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071505 4775 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071514 4775 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071522 4775 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071532 4775 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071542 4775 flags.go:64] FLAG: --contention-profiling="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071551 4775 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071560 4775 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071569 4775 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071578 4775 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071589 4775 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071598 4775 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071607 4775 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071616 4775 flags.go:64] FLAG: --enable-load-reader="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071625 4775 flags.go:64] FLAG: --enable-server="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071634 4775 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071650 4775 flags.go:64] FLAG: --event-burst="100" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071659 4775 flags.go:64] FLAG: --event-qps="50" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071669 4775 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071678 4775 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071687 4775 flags.go:64] FLAG: --eviction-hard="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071706 4775 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071741 4775 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071750 4775 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071763 4775 flags.go:64] FLAG: --eviction-soft="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071772 4775 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071781 4775 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071790 4775 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071798 4775 flags.go:64] FLAG: --experimental-mounter-path="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071809 4775 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071818 4775 flags.go:64] FLAG: --fail-swap-on="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071827 4775 flags.go:64] FLAG: --feature-gates="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071838 4775 flags.go:64] FLAG: --file-check-frequency="20s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071847 4775 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071856 4775 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071865 4775 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071874 4775 flags.go:64] FLAG: --healthz-port="10248" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071885 4775 flags.go:64] FLAG: --help="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071894 4775 flags.go:64] FLAG: --hostname-override="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071903 4775 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071912 4775 flags.go:64] FLAG: --http-check-frequency="20s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071921 4775 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071930 4775 flags.go:64] FLAG: --image-credential-provider-config="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071939 4775 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071948 4775 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071957 4775 flags.go:64] FLAG: --image-service-endpoint="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071965 4775 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071974 4775 flags.go:64] FLAG: --kube-api-burst="100" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071983 4775 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.071993 4775 flags.go:64] FLAG: --kube-api-qps="50" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072001 4775 flags.go:64] FLAG: --kube-reserved="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072010 4775 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072019 4775 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072029 4775 flags.go:64] FLAG: --kubelet-cgroups="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072037 4775 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072046 4775 flags.go:64] FLAG: --lock-file="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072055 4775 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072064 4775 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072073 4775 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072087 4775 flags.go:64] FLAG: --log-json-split-stream="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072097 4775 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072106 4775 flags.go:64] FLAG: --log-text-split-stream="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072115 4775 flags.go:64] FLAG: --logging-format="text" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072124 4775 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072133 4775 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072142 4775 flags.go:64] FLAG: --manifest-url="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072151 4775 flags.go:64] FLAG: --manifest-url-header="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072163 4775 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072172 4775 flags.go:64] FLAG: --max-open-files="1000000" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072183 4775 flags.go:64] FLAG: --max-pods="110" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072192 4775 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072201 4775 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072210 4775 flags.go:64] FLAG: --memory-manager-policy="None" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072220 4775 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072229 4775 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072238 4775 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072246 4775 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072267 4775 flags.go:64] FLAG: --node-status-max-images="50" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072277 4775 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072286 4775 flags.go:64] FLAG: --oom-score-adj="-999" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072295 4775 flags.go:64] FLAG: --pod-cidr="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072304 4775 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072317 4775 flags.go:64] FLAG: --pod-manifest-path="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072326 4775 flags.go:64] FLAG: --pod-max-pids="-1" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072335 4775 flags.go:64] FLAG: --pods-per-core="0" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072344 4775 flags.go:64] FLAG: --port="10250" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072353 4775 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072362 4775 flags.go:64] FLAG: --provider-id="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072371 4775 flags.go:64] FLAG: --qos-reserved="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072380 4775 flags.go:64] FLAG: --read-only-port="10255" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072389 4775 flags.go:64] FLAG: --register-node="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072397 4775 flags.go:64] FLAG: --register-schedulable="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072406 4775 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072422 4775 flags.go:64] FLAG: --registry-burst="10" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072430 4775 flags.go:64] FLAG: --registry-qps="5" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072439 4775 flags.go:64] FLAG: --reserved-cpus="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072457 4775 flags.go:64] FLAG: --reserved-memory="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072475 4775 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072485 4775 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072494 4775 flags.go:64] FLAG: --rotate-certificates="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072503 4775 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072511 4775 flags.go:64] FLAG: --runonce="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072520 4775 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072530 4775 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072540 4775 flags.go:64] FLAG: --seccomp-default="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072549 4775 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072557 4775 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072566 4775 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072575 4775 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072585 4775 flags.go:64] FLAG: --storage-driver-password="root" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072593 4775 flags.go:64] FLAG: --storage-driver-secure="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072602 4775 flags.go:64] FLAG: --storage-driver-table="stats" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072611 4775 flags.go:64] FLAG: --storage-driver-user="root" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072619 4775 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072629 4775 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072638 4775 flags.go:64] FLAG: --system-cgroups="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072647 4775 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072661 4775 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072669 4775 flags.go:64] FLAG: --tls-cert-file="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072678 4775 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072690 4775 flags.go:64] FLAG: --tls-min-version="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072699 4775 flags.go:64] FLAG: --tls-private-key-file="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072708 4775 flags.go:64] FLAG: --topology-manager-policy="none" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072740 4775 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072750 4775 flags.go:64] FLAG: --topology-manager-scope="container" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072760 4775 flags.go:64] FLAG: --v="2" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072771 4775 flags.go:64] FLAG: --version="false" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072782 4775 flags.go:64] FLAG: --vmodule="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072793 4775 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.072802 4775 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073011 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073021 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073032 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073041 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073050 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073058 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073067 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073076 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073087 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073097 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073105 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073114 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073122 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073131 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073139 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073146 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073157 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073166 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073177 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073185 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073193 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073201 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073211 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073219 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073227 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073235 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073242 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073264 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073271 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073279 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073286 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073295 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073303 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073310 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073318 4775 feature_gate.go:330] unrecognized feature gate: Example Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073325 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073338 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073346 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073355 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073363 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073371 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073379 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073387 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073397 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073407 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073417 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073426 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073435 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073444 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073453 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073462 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073471 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073480 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073488 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073497 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073505 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073512 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073521 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073528 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073538 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073546 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073553 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073561 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073569 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073577 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073585 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073592 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073599 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073610 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073618 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.073625 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.073650 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.085494 4775 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.085540 4775 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085663 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085677 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085687 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085697 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085705 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085748 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085761 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085772 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085781 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085791 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085800 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085807 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085815 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085823 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085831 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085839 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085846 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085854 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085862 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085870 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085877 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085885 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085893 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085902 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085910 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085918 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085925 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085933 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085942 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085951 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085959 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085968 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085976 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085984 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.085993 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086001 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086009 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086017 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086024 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086032 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086043 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086054 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086065 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086075 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086084 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086093 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086101 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086109 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086117 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086125 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086132 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086140 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086150 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086161 4775 feature_gate.go:330] unrecognized feature gate: Example Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086171 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086183 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086192 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086200 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086208 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086217 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086226 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086234 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086243 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086251 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086260 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086268 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086276 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086286 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086295 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086304 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086312 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.086325 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086573 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086587 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086596 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086604 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086614 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086625 4775 feature_gate.go:330] unrecognized feature gate: Example Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086634 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086643 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086652 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086661 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086669 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086677 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086686 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086694 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086703 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086711 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086753 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086763 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086773 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086780 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086789 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086796 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086807 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086816 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086824 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086833 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086841 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086849 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086857 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086865 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086872 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086880 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086888 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086896 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086905 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086913 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086920 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086928 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086936 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086944 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086952 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086960 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086968 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086975 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086984 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086991 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.086999 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087007 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087015 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087023 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087030 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087038 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087045 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087053 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087061 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087069 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087077 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087084 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087092 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087100 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087109 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087118 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087126 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087133 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087141 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087149 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087156 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087164 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087172 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087182 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.087194 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.087206 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.087417 4775 server.go:940] "Client rotation is on, will bootstrap in background" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.092902 4775 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.093020 4775 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.094751 4775 server.go:997] "Starting client certificate rotation" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.094802 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.095062 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-15 21:01:30.36436471 +0000 UTC Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.095202 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 470h36m13.269169109s for next certificate rotation Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.125400 4775 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.130777 4775 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.148797 4775 log.go:25] "Validated CRI v1 runtime API" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.188926 4775 log.go:25] "Validated CRI v1 image API" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.191405 4775 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.199652 4775 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-26-06-21-00-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.199794 4775 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:46 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:42 fsType:tmpfs blockSize:0}] Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.225754 4775 manager.go:217] Machine: {Timestamp:2025-11-26 06:25:17.223316981 +0000 UTC m=+0.584621003 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:7c9a1c38-84d7-4412-b12e-7045594af10c BootID:ca7ec27e-1a29-480c-b7d0-2bfb73b424b6 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:46 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:42 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:3b:fa:fd Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:3b:fa:fd Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:30:93:42 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c5:c4:15 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:7f:39:b0 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:20:96:e8 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:16:da:c6:79:64:43 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:81:47:db:48:43 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.226308 4775 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.226529 4775 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.228567 4775 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.228904 4775 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.228961 4775 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.229286 4775 topology_manager.go:138] "Creating topology manager with none policy" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.229305 4775 container_manager_linux.go:303] "Creating device plugin manager" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.229992 4775 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.230045 4775 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.230383 4775 state_mem.go:36] "Initialized new in-memory state store" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.230537 4775 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.235470 4775 kubelet.go:418] "Attempting to sync node with API server" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.235508 4775 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.235542 4775 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.235564 4775 kubelet.go:324] "Adding apiserver pod source" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.235582 4775 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.239968 4775 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.241394 4775 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.243176 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.243249 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.243357 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.243287 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.244173 4775 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245683 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245754 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245770 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245784 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245811 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245827 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245841 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245864 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245882 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245905 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245923 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245937 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.245985 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.246639 4775 server.go:1280] "Started kubelet" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.246679 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.246899 4775 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.247134 4775 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.247461 4775 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 26 06:25:17 crc systemd[1]: Started Kubernetes Kubelet. Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.253965 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.254056 4775 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.254417 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 23:30:49.754359703 +0000 UTC Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.254538 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 641h5m32.499829241s for next certificate rotation Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.254828 4775 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.254868 4775 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.254910 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.255030 4775 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.255631 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.256491 4775 server.go:460] "Adding debug handlers to kubelet server" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.256568 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.257502 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.257847 4775 factory.go:55] Registering systemd factory Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.256436 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b7a73e623541e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 06:25:17.246583838 +0000 UTC m=+0.607887830,LastTimestamp:2025-11-26 06:25:17.246583838 +0000 UTC m=+0.607887830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.257877 4775 factory.go:221] Registration of the systemd container factory successfully Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.261149 4775 factory.go:153] Registering CRI-O factory Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.261192 4775 factory.go:221] Registration of the crio container factory successfully Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.261312 4775 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.261351 4775 factory.go:103] Registering Raw factory Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.261383 4775 manager.go:1196] Started watching for new ooms in manager Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.262343 4775 manager.go:319] Starting recovery of all containers Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.269978 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270059 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270088 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270115 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270139 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270163 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270191 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270214 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270241 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270267 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270293 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270319 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270343 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270373 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270399 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270515 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270541 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270564 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270588 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270612 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270636 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270662 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270687 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270746 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270838 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270870 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270905 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270936 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270961 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.270999 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271036 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271065 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271090 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271124 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271157 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271193 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271220 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271244 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271270 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271296 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271319 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271342 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271366 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271390 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271412 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271435 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271457 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271486 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271510 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271535 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271562 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271587 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271623 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271651 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271676 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271702 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271763 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271790 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271815 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271838 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271865 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271889 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271915 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.271941 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272017 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272042 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272069 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272094 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272117 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272141 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272165 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272189 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272214 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272237 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272261 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272300 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272327 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272351 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272375 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272405 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272428 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272453 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.272477 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274864 4775 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274912 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274933 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274948 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274961 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274976 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.274991 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275004 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275018 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275031 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275044 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275058 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275071 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275084 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275097 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275111 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275124 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275139 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275152 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275166 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275183 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275201 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275237 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275254 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275269 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275283 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275299 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275315 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275331 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275346 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275360 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275373 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275388 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275402 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275414 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275426 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275438 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275462 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275474 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275487 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275500 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275512 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275526 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275538 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275551 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275565 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275579 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275590 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275603 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275617 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275629 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275641 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275656 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275668 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275682 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275694 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275706 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275735 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275748 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275759 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275773 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275786 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275797 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275812 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275824 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275838 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275851 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275863 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275874 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275886 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275898 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275913 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275927 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275939 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275952 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275964 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275987 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.275999 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276012 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276024 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276036 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276048 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276061 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276073 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276084 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276095 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276107 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276120 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276132 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276145 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276156 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276168 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276181 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276194 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276206 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276217 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276230 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276242 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276253 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276266 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276277 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276289 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276303 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276316 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276330 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276342 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276354 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276365 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276380 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276392 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276405 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276417 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276430 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276441 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276454 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276466 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276480 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276492 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276503 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276515 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276526 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276538 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276549 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276560 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276574 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276586 4775 reconstruct.go:97] "Volume reconstruction finished" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.276595 4775 reconciler.go:26] "Reconciler: start to sync state" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.302072 4775 manager.go:324] Recovery completed Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.318682 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.320844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.320906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.320925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.323072 4775 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.325123 4775 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.325156 4775 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.325182 4775 state_mem.go:36] "Initialized new in-memory state store" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.326315 4775 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.326361 4775 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.326392 4775 kubelet.go:2335] "Starting kubelet main sync loop" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.326512 4775 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.329056 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.329158 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.349821 4775 policy_none.go:49] "None policy: Start" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.350950 4775 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.350997 4775 state_mem.go:35] "Initializing new in-memory state store" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.355077 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.417292 4775 manager.go:334] "Starting Device Plugin manager" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.417585 4775 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.417615 4775 server.go:79] "Starting device plugin registration server" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.418211 4775 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.418239 4775 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.419378 4775 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.419482 4775 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.419495 4775 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.426683 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.426693 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.426815 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.427999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.428067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.428093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.428372 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.428551 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.428615 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.429900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.429925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.429945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.429956 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.429930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.430053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.430235 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.430344 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.430389 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.431643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.431671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.431747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.431779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.431760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.431859 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.432050 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.432271 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.432336 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433234 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433411 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433464 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433918 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433953 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.433963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434110 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434137 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434751 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.434851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.456705 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.478974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.479053 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.479086 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.479153 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.483032 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.484745 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485175 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485411 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485547 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485627 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.485780 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.518658 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.520400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.520442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.520454 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.520487 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.521182 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588034 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588139 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588189 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588319 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588446 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588446 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588318 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588517 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588542 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588528 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588596 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588609 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588759 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588867 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588951 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.588960 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589011 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589124 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589190 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589210 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.589243 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.722155 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.723685 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.723829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.723866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.723914 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.724468 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.769392 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.780397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.803284 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.817816 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: I1126 06:25:17.825072 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.832360 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-3041656797e791e2e955f480dbfe801ff0996bcdbc62d96e795e25f7f9149783 WatchSource:0}: Error finding container 3041656797e791e2e955f480dbfe801ff0996bcdbc62d96e795e25f7f9149783: Status 404 returned error can't find the container with id 3041656797e791e2e955f480dbfe801ff0996bcdbc62d96e795e25f7f9149783 Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.843358 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ad603af5b02050cafc20bc8c3b6e2a4572aa9903f21a8aa5356bac040d969fdb WatchSource:0}: Error finding container ad603af5b02050cafc20bc8c3b6e2a4572aa9903f21a8aa5356bac040d969fdb: Status 404 returned error can't find the container with id ad603af5b02050cafc20bc8c3b6e2a4572aa9903f21a8aa5356bac040d969fdb Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.853081 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3b4dce88b27c3e7231ba19821a5932cfe75abb6eadc7be912da590c554db010d WatchSource:0}: Error finding container 3b4dce88b27c3e7231ba19821a5932cfe75abb6eadc7be912da590c554db010d: Status 404 returned error can't find the container with id 3b4dce88b27c3e7231ba19821a5932cfe75abb6eadc7be912da590c554db010d Nov 26 06:25:17 crc kubenswrapper[4775]: E1126 06:25:17.858200 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Nov 26 06:25:17 crc kubenswrapper[4775]: W1126 06:25:17.861290 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-d3fb84903efb2b153409db3f34edc2d610133813e883962eda53651ffb6096ea WatchSource:0}: Error finding container d3fb84903efb2b153409db3f34edc2d610133813e883962eda53651ffb6096ea: Status 404 returned error can't find the container with id d3fb84903efb2b153409db3f34edc2d610133813e883962eda53651ffb6096ea Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.124628 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.126954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.127028 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.127102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.127175 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.127940 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 06:25:18 crc kubenswrapper[4775]: W1126 06:25:18.154830 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.154962 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:18 crc kubenswrapper[4775]: W1126 06:25:18.187429 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.187519 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.248189 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.331755 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d623d1eb6ca46563ef719b7a2d5be5816d69d109ab69c02cbe3474dcfc51636d"} Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.332912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d3fb84903efb2b153409db3f34edc2d610133813e883962eda53651ffb6096ea"} Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.334091 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3b4dce88b27c3e7231ba19821a5932cfe75abb6eadc7be912da590c554db010d"} Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.335414 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ad603af5b02050cafc20bc8c3b6e2a4572aa9903f21a8aa5356bac040d969fdb"} Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.336126 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3041656797e791e2e955f480dbfe801ff0996bcdbc62d96e795e25f7f9149783"} Nov 26 06:25:18 crc kubenswrapper[4775]: W1126 06:25:18.441694 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.442293 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:18 crc kubenswrapper[4775]: W1126 06:25:18.550636 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.550735 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.659218 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.929070 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.930815 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.930884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.930905 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:18 crc kubenswrapper[4775]: I1126 06:25:18.930948 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:18 crc kubenswrapper[4775]: E1126 06:25:18.931607 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.248378 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.341483 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.341535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.341551 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.341570 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.341684 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.342965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.342990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.342998 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.345738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.345889 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.345589 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96" exitCode=0 Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.347539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.347575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.347586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.348527 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1a3c70d0749f7ce3261a31f6900a8f8beb0eb4d1dd3bc80e664cb811c6e2be79" exitCode=0 Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.348650 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.348850 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1a3c70d0749f7ce3261a31f6900a8f8beb0eb4d1dd3bc80e664cb811c6e2be79"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349438 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.349848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.352275 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10" exitCode=0 Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.352355 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.352509 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.353943 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.353976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.353993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.356671 4775 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b" exitCode=0 Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.356747 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.356743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b"} Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.358620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.358642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:19 crc kubenswrapper[4775]: I1126 06:25:19.358653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:20 crc kubenswrapper[4775]: W1126 06:25:20.176156 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:20 crc kubenswrapper[4775]: E1126 06:25:20.176291 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.248999 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:20 crc kubenswrapper[4775]: E1126 06:25:20.260419 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.370481 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.370544 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.370564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.370581 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.372071 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e377bd9acdbc1222ab56746d9f9f017243671347c15a0058a9881599dd3f4ab2" exitCode=0 Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.372138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e377bd9acdbc1222ab56746d9f9f017243671347c15a0058a9881599dd3f4ab2"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.372174 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.373125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.373169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.373183 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.374733 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1d9cb9007f907be0090f6b917c6a58f654971b43b1ff4d278d5787f1852c2ebf"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.374797 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.375757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.375786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.375797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.385862 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.385900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.385912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7"} Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.385933 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.385948 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.386963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.387004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.387029 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.387913 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.387944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.387977 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.531780 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.535105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.535135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.535147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:20 crc kubenswrapper[4775]: I1126 06:25:20.535172 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:20 crc kubenswrapper[4775]: E1126 06:25:20.536014 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 06:25:20 crc kubenswrapper[4775]: W1126 06:25:20.589052 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 06:25:20 crc kubenswrapper[4775]: E1126 06:25:20.589319 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 06:25:20 crc kubenswrapper[4775]: E1126 06:25:20.610377 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b7a73e623541e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 06:25:17.246583838 +0000 UTC m=+0.607887830,LastTimestamp:2025-11-26 06:25:17.246583838 +0000 UTC m=+0.607887830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.392454 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae"} Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.393381 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.395014 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.395043 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.395053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.395985 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="60405c739c33461cccae02ec424977bfd37af0f9949174be24264fa98bf7e187" exitCode=0 Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396055 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396104 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396038 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"60405c739c33461cccae02ec424977bfd37af0f9949174be24264fa98bf7e187"} Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396301 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396365 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396907 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396883 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.396975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.397781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.397835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.397856 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.887229 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.887457 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.888871 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.888928 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.888952 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:21 crc kubenswrapper[4775]: I1126 06:25:21.932359 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.405841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8da2deca0116cc9dc59aa8f63853173ac44e5b14930aa01ebe8fbec42e6f6a2d"} Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.405897 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.405914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a27d04810f38053c40b22febf525ced1ab878dfd496a9ae51c60558e3c611a19"} Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.405944 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3e2811bdd7a1cbf68fc5a7571805837c30dd15f653f0203446d9387d4f7d9770"} Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.406109 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.408534 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.408910 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.409122 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.409181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.412868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.412944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:22 crc kubenswrapper[4775]: I1126 06:25:22.413000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.416047 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8d31054d8b35eeb476c62d7e35e0ca196aec8eb782e74f4ac9d481b20cc7d421"} Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.416124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dcf3f13ae870008d0232c731f82dc10cbe3d8f366d96e5d072e334e157368b3c"} Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.416200 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.416885 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.417637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.417688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.417707 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.418267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.418304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.418319 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.432113 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.736897 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.738533 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.738583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.738603 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:23 crc kubenswrapper[4775]: I1126 06:25:23.738644 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.049984 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.050284 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.051848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.051915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.051934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.059988 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.419148 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.419286 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.419354 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.420474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.420541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.420559 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.421256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.421290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.421303 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.421410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.421449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:24 crc kubenswrapper[4775]: I1126 06:25:24.421468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:26 crc kubenswrapper[4775]: I1126 06:25:26.943653 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:26 crc kubenswrapper[4775]: I1126 06:25:26.943987 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:26 crc kubenswrapper[4775]: I1126 06:25:26.945577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:26 crc kubenswrapper[4775]: I1126 06:25:26.945771 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:26 crc kubenswrapper[4775]: I1126 06:25:26.945802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:27 crc kubenswrapper[4775]: E1126 06:25:27.426841 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.349626 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.349762 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.349995 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.350066 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.352041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.352101 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.352104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.352188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.352207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.352123 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.356579 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.429551 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.430869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.430919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:28 crc kubenswrapper[4775]: I1126 06:25:28.430936 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:29 crc kubenswrapper[4775]: I1126 06:25:29.944130 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 26 06:25:29 crc kubenswrapper[4775]: I1126 06:25:29.944224 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 06:25:30 crc kubenswrapper[4775]: I1126 06:25:30.053955 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 26 06:25:30 crc kubenswrapper[4775]: I1126 06:25:30.054019 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 26 06:25:31 crc kubenswrapper[4775]: W1126 06:25:31.162553 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.162680 4775 trace.go:236] Trace[942976248]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 06:25:21.161) (total time: 10001ms): Nov 26 06:25:31 crc kubenswrapper[4775]: Trace[942976248]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:25:31.162) Nov 26 06:25:31 crc kubenswrapper[4775]: Trace[942976248]: [10.001550314s] [10.001550314s] END Nov 26 06:25:31 crc kubenswrapper[4775]: E1126 06:25:31.162751 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 26 06:25:31 crc kubenswrapper[4775]: W1126 06:25:31.222640 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.222796 4775 trace.go:236] Trace[1218458085]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 06:25:21.221) (total time: 10001ms): Nov 26 06:25:31 crc kubenswrapper[4775]: Trace[1218458085]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:25:31.222) Nov 26 06:25:31 crc kubenswrapper[4775]: Trace[1218458085]: [10.00167754s] [10.00167754s] END Nov 26 06:25:31 crc kubenswrapper[4775]: E1126 06:25:31.222827 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.248232 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.625327 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.625385 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.633956 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.634278 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.941700 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]log ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]etcd ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/generic-apiserver-start-informers ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-filter ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-apiextensions-informers ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-apiextensions-controllers ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/crd-informer-synced ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-system-namespaces-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 26 06:25:31 crc kubenswrapper[4775]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 26 06:25:31 crc kubenswrapper[4775]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/bootstrap-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/start-kube-aggregator-informers ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-registration-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-discovery-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]autoregister-completion ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-openapi-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 26 06:25:31 crc kubenswrapper[4775]: livez check failed Nov 26 06:25:31 crc kubenswrapper[4775]: I1126 06:25:31.941856 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:25:32 crc kubenswrapper[4775]: I1126 06:25:32.571321 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 26 06:25:32 crc kubenswrapper[4775]: I1126 06:25:32.571640 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:32 crc kubenswrapper[4775]: I1126 06:25:32.573202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:32 crc kubenswrapper[4775]: I1126 06:25:32.573311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:32 crc kubenswrapper[4775]: I1126 06:25:32.573398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:32 crc kubenswrapper[4775]: I1126 06:25:32.662251 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 26 06:25:33 crc kubenswrapper[4775]: I1126 06:25:33.374382 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 26 06:25:33 crc kubenswrapper[4775]: I1126 06:25:33.444285 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:33 crc kubenswrapper[4775]: I1126 06:25:33.446214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:33 crc kubenswrapper[4775]: I1126 06:25:33.446406 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:33 crc kubenswrapper[4775]: I1126 06:25:33.446542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:34 crc kubenswrapper[4775]: I1126 06:25:34.447907 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:34 crc kubenswrapper[4775]: I1126 06:25:34.450198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:34 crc kubenswrapper[4775]: I1126 06:25:34.450260 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:34 crc kubenswrapper[4775]: I1126 06:25:34.450277 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:34 crc kubenswrapper[4775]: I1126 06:25:34.538244 4775 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 26 06:25:35 crc kubenswrapper[4775]: I1126 06:25:35.236995 4775 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 26 06:25:36 crc kubenswrapper[4775]: E1126 06:25:36.617469 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.621782 4775 trace.go:236] Trace[1487827185]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 06:25:24.932) (total time: 11689ms): Nov 26 06:25:36 crc kubenswrapper[4775]: Trace[1487827185]: ---"Objects listed" error: 11689ms (06:25:36.621) Nov 26 06:25:36 crc kubenswrapper[4775]: Trace[1487827185]: [11.689511386s] [11.689511386s] END Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.621815 4775 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.622127 4775 trace.go:236] Trace[726273455]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 06:25:25.267) (total time: 11354ms): Nov 26 06:25:36 crc kubenswrapper[4775]: Trace[726273455]: ---"Objects listed" error: 11354ms (06:25:36.622) Nov 26 06:25:36 crc kubenswrapper[4775]: Trace[726273455]: [11.354604636s] [11.354604636s] END Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.622173 4775 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 26 06:25:36 crc kubenswrapper[4775]: E1126 06:25:36.623273 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.626565 4775 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.692330 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50232->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.692410 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:50232->192.168.126.11:17697: read: connection reset by peer" Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.943689 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.944857 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.944961 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 26 06:25:36 crc kubenswrapper[4775]: I1126 06:25:36.949391 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.034194 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.043945 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.248533 4775 apiserver.go:52] "Watching apiserver" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.250836 4775 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.251962 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-dns/node-resolver-t5r7v"] Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.253078 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.253466 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.253861 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.253941 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.254195 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.254249 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.254316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.254699 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.254927 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.254963 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.255010 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.255487 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.255946 4775 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.257366 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258067 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258094 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258298 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258456 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258491 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258599 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258641 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258787 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.258892 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.269304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.279832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.294561 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.305488 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.314428 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.324307 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331164 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331204 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331222 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331243 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331282 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331567 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331600 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331626 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331650 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331672 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331743 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331766 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331780 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331811 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331828 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331845 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331879 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331894 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331910 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.331977 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332052 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332076 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332269 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332298 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332282 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332300 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332349 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332505 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332510 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332513 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332629 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332692 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332729 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332706 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332849 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332912 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332969 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332978 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333109 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333124 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333136 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333155 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.332210 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333253 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333270 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333282 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333286 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333333 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333358 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333379 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333398 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333417 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333437 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333454 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333471 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333505 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333525 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333544 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333560 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333579 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333606 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333895 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.333912 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334091 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334139 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334177 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334197 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334221 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334237 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334255 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334273 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334291 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334300 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334336 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334353 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334874 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334985 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335006 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335025 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335042 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335061 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335077 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335093 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335169 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335187 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335204 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335236 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335254 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335271 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335322 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335385 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335406 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335471 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335538 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335557 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335576 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335602 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335641 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335658 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335675 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335734 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335752 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335768 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335786 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335803 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335836 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335856 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335873 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335890 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335907 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335924 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335945 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335962 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336000 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336018 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336050 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336066 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336081 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336097 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336114 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336132 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336148 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336164 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336181 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336197 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336213 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336230 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336247 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336958 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337007 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337037 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337062 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337086 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337106 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337128 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337150 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340294 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340353 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340377 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340400 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340421 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340445 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340466 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340485 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340508 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340572 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340602 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340642 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340664 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340681 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340704 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340738 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340759 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340781 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340805 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340826 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340846 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340866 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340888 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340909 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341023 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341050 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341072 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341092 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341111 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341132 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341152 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341170 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341202 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341240 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341282 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341305 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341325 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341345 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341366 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341388 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341409 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341449 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341470 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341488 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341507 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341528 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341548 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341565 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341585 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341607 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341624 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341646 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341666 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341685 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341703 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341749 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341816 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341833 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341855 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341878 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341956 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342006 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342030 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342051 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn4jr\" (UniqueName: \"kubernetes.io/projected/5458a69d-5519-4863-baf9-e549e5e52b78-kube-api-access-bn4jr\") pod \"node-resolver-t5r7v\" (UID: \"5458a69d-5519-4863-baf9-e549e5e52b78\") " pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342096 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342120 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342141 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342209 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342230 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5458a69d-5519-4863-baf9-e549e5e52b78-hosts-file\") pod \"node-resolver-t5r7v\" (UID: \"5458a69d-5519-4863-baf9-e549e5e52b78\") " pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342253 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342322 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342342 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342622 4775 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342635 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342646 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342660 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342670 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342680 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342691 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342706 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342729 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342739 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342749 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342762 4775 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342772 4775 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342782 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342916 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342927 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342939 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342951 4775 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342964 4775 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342973 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342983 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342993 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343007 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343017 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343028 4775 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343041 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343050 4775 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343340 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334469 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334698 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334758 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334767 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.334842 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335243 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335268 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335588 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335864 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.335917 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.336084 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.337589 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.338129 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.338172 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.338333 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.338357 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.338891 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.339024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.339668 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.345457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.345691 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.345889 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.345998 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346309 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.339967 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340011 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340358 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.348622 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.348812 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.348963 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.349055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340916 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340936 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341167 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341242 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341326 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341362 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.341975 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342054 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350529 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350540 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342415 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342220 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342557 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342584 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.349897 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343335 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343410 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343764 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343919 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350694 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343925 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343994 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.344123 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.344249 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.343324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.344481 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.344480 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346319 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347437 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347455 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346569 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.344353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347478 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346959 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347750 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347810 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347827 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.347910 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.348102 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:25:37.848056857 +0000 UTC m=+21.209360839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.339800 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.349415 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.340362 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.349845 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.349883 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350066 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350557 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.342918 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.351337 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350922 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.351051 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.351189 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.350101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.351405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.346415 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.351481 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353105 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353151 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353260 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353409 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353702 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353874 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353909 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.353933 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.354015 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.354125 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.354441 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.354541 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.354304 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.354756 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.355071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.355326 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.355398 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.355581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.356164 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.356443 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.356486 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.356611 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.357416 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.357732 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.357853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.358384 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.358488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.358010 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.358460 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.358935 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.359071 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.359133 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:37.859108472 +0000 UTC m=+21.220412424 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.359219 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.359438 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.359514 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.359541 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.359756 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.359790 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.360076 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.359882 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.360283 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.360498 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.360540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.360825 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:37.860806285 +0000 UTC m=+21.222110357 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.361794 4775 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.362405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.364402 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.364662 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.365092 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.365874 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.365927 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.367472 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.368258 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.369251 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.369815 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.371298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.372513 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.372894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.372994 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.373237 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.373522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.374117 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.374136 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.374147 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.374199 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:37.87418338 +0000 UTC m=+21.235487332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.374952 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.375097 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.376155 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.376249 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.376516 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.377259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.377392 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.381766 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.383028 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.385490 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.385513 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.385524 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.385567 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:37.885551672 +0000 UTC m=+21.246855624 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.385827 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.385835 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.386382 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.386780 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.387908 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389140 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389253 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389417 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389239 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389790 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389874 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.389957 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.390801 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.391802 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.392993 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.394006 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.405695 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.408834 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.416493 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.418079 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.421251 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.427273 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.455975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn4jr\" (UniqueName: \"kubernetes.io/projected/5458a69d-5519-4863-baf9-e549e5e52b78-kube-api-access-bn4jr\") pod \"node-resolver-t5r7v\" (UID: \"5458a69d-5519-4863-baf9-e549e5e52b78\") " pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456073 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456120 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5458a69d-5519-4863-baf9-e549e5e52b78-hosts-file\") pod \"node-resolver-t5r7v\" (UID: \"5458a69d-5519-4863-baf9-e549e5e52b78\") " pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456217 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456240 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456259 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456277 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456294 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456311 4775 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456328 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456345 4775 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456362 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456379 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456396 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456412 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456431 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456448 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456465 4775 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456483 4775 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456501 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456518 4775 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456534 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456552 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456569 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456587 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456604 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456620 4775 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456638 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456654 4775 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456671 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456688 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456705 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456754 4775 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456771 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456789 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456809 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456827 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456844 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456861 4775 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456878 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456895 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456911 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456930 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456947 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456964 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456980 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.456997 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457015 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457033 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457050 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457068 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457141 4775 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457162 4775 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457183 4775 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457200 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457217 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457234 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457251 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457268 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457286 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457312 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457329 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457346 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457364 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457381 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457399 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457416 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457432 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457448 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457466 4775 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457483 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457499 4775 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457516 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457535 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457552 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457569 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457588 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457605 4775 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457623 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457639 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457656 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457673 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457690 4775 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457707 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457752 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457771 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457787 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457803 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457820 4775 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457836 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457854 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457871 4775 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457891 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457910 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457929 4775 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457945 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457961 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457977 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.457997 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458013 4775 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458030 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458047 4775 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458064 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458080 4775 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458097 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458114 4775 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458130 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458146 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458162 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458180 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458197 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458215 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458232 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458249 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458266 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458283 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458315 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458332 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458349 4775 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458371 4775 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458390 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458407 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458424 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458442 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458459 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458514 4775 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458532 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458588 4775 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458606 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458624 4775 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458642 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458796 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458815 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458908 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458927 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458943 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458960 4775 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458976 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.458992 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459009 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459026 4775 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459070 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459081 4775 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459092 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459101 4775 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459110 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459107 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5458a69d-5519-4863-baf9-e549e5e52b78-hosts-file\") pod \"node-resolver-t5r7v\" (UID: \"5458a69d-5519-4863-baf9-e549e5e52b78\") " pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459120 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459207 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459226 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459244 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459263 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459281 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459298 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459314 4775 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459330 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459346 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459368 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459385 4775 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459401 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459418 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459436 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459454 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459470 4775 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459488 4775 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459504 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459521 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459537 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459554 4775 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459570 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459587 4775 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.459635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.461682 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.463489 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae" exitCode=255 Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.463574 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae"} Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.467004 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.468843 4775 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.469150 4775 scope.go:117] "RemoveContainer" containerID="81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.472200 4775 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.476596 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.480692 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn4jr\" (UniqueName: \"kubernetes.io/projected/5458a69d-5519-4863-baf9-e549e5e52b78-kube-api-access-bn4jr\") pod \"node-resolver-t5r7v\" (UID: \"5458a69d-5519-4863-baf9-e549e5e52b78\") " pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.497851 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.512385 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.526012 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.547468 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.560927 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.571761 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.573975 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-t5r7v" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.583938 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.587062 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.597605 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.596304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: W1126 06:25:37.605788 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d903149f5d6d090f01a1d9c061096907e90524df7d4301044df7e0409924b732 WatchSource:0}: Error finding container d903149f5d6d090f01a1d9c061096907e90524df7d4301044df7e0409924b732: Status 404 returned error can't find the container with id d903149f5d6d090f01a1d9c061096907e90524df7d4301044df7e0409924b732 Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.606220 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.608797 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.619474 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: W1126 06:25:37.626777 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-25e739cdc7413b891671f4acb013457a311a7b85123aa8c53caae7aacf7fadf0 WatchSource:0}: Error finding container 25e739cdc7413b891671f4acb013457a311a7b85123aa8c53caae7aacf7fadf0: Status 404 returned error can't find the container with id 25e739cdc7413b891671f4acb013457a311a7b85123aa8c53caae7aacf7fadf0 Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.628569 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.642545 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.863198 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.863291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.863338 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:25:38.863318269 +0000 UTC m=+22.224622221 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.863362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.863418 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.863440 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.863460 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:38.863451062 +0000 UTC m=+22.224755014 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.863474 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:38.863466963 +0000 UTC m=+22.224770915 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.964416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.964467 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964575 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964590 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964601 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964648 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:38.964635347 +0000 UTC m=+22.325939299 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964777 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964822 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964841 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: E1126 06:25:37.964932 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:38.964906864 +0000 UTC m=+22.326210846 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.972374 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-k9pcc"] Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.972962 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-p7z2d"] Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.973193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k9pcc" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.973472 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.974897 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.975248 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.975276 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.975501 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.975805 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.975832 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.975935 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.976229 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.977868 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.978112 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 26 06:25:37 crc kubenswrapper[4775]: I1126 06:25:37.992838 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.000328 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.011785 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.022025 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.036235 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.053669 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065759 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-os-release\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-hostroot\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065812 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-cni-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/687d9603-b248-4ad7-95d6-a23c110afaf1-rootfs\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065855 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-daemon-config\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065874 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/687d9603-b248-4ad7-95d6-a23c110afaf1-mcd-auth-proxy-config\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065898 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-multus-certs\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065918 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srhx4\" (UniqueName: \"kubernetes.io/projected/4181bfb8-478b-4b30-a2cb-744da1010aeb-kube-api-access-srhx4\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065934 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-cni-bin\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065948 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-kubelet\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-etc-kubernetes\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065977 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-cnibin\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.065998 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-system-cni-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066014 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4181bfb8-478b-4b30-a2cb-744da1010aeb-cni-binary-copy\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066030 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-socket-dir-parent\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066045 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-k8s-cni-cncf-io\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066083 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/687d9603-b248-4ad7-95d6-a23c110afaf1-proxy-tls\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066108 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx67z\" (UniqueName: \"kubernetes.io/projected/687d9603-b248-4ad7-95d6-a23c110afaf1-kube-api-access-dx67z\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066149 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-netns\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066163 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-cni-multus\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.066178 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-conf-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.069966 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.086263 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.099105 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.116281 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.126075 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.138843 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.151295 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.163263 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.166875 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-k8s-cni-cncf-io\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.166953 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-k8s-cni-cncf-io\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.166986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/687d9603-b248-4ad7-95d6-a23c110afaf1-proxy-tls\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx67z\" (UniqueName: \"kubernetes.io/projected/687d9603-b248-4ad7-95d6-a23c110afaf1-kube-api-access-dx67z\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-netns\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167062 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-cni-multus\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-conf-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167118 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-netns\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167300 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-cni-multus\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-os-release\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167358 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-hostroot\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-os-release\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167618 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/687d9603-b248-4ad7-95d6-a23c110afaf1-rootfs\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-cni-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-conf-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167670 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-daemon-config\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167645 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/687d9603-b248-4ad7-95d6-a23c110afaf1-rootfs\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167793 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-cni-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167822 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-multus-certs\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167837 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srhx4\" (UniqueName: \"kubernetes.io/projected/4181bfb8-478b-4b30-a2cb-744da1010aeb-kube-api-access-srhx4\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/687d9603-b248-4ad7-95d6-a23c110afaf1-mcd-auth-proxy-config\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168269 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-cni-bin\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168330 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-cni-bin\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167971 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-hostroot\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168294 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-kubelet\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.167902 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-run-multus-certs\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-host-var-lib-kubelet\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-etc-kubernetes\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168486 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-etc-kubernetes\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168494 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-system-cni-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168533 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-cnibin\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168559 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4181bfb8-478b-4b30-a2cb-744da1010aeb-cni-binary-copy\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168584 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-socket-dir-parent\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-cnibin\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168643 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-socket-dir-parent\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4181bfb8-478b-4b30-a2cb-744da1010aeb-multus-daemon-config\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.168658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/687d9603-b248-4ad7-95d6-a23c110afaf1-mcd-auth-proxy-config\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.169003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4181bfb8-478b-4b30-a2cb-744da1010aeb-cni-binary-copy\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.169049 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4181bfb8-478b-4b30-a2cb-744da1010aeb-system-cni-dir\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.172159 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/687d9603-b248-4ad7-95d6-a23c110afaf1-proxy-tls\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.178820 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.190914 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx67z\" (UniqueName: \"kubernetes.io/projected/687d9603-b248-4ad7-95d6-a23c110afaf1-kube-api-access-dx67z\") pod \"machine-config-daemon-p7z2d\" (UID: \"687d9603-b248-4ad7-95d6-a23c110afaf1\") " pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.193153 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srhx4\" (UniqueName: \"kubernetes.io/projected/4181bfb8-478b-4b30-a2cb-744da1010aeb-kube-api-access-srhx4\") pod \"multus-k9pcc\" (UID: \"4181bfb8-478b-4b30-a2cb-744da1010aeb\") " pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.196066 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.205054 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.212991 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.223159 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.233059 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.247083 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.285549 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k9pcc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.289429 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:25:38 crc kubenswrapper[4775]: W1126 06:25:38.305980 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4181bfb8_478b_4b30_a2cb_744da1010aeb.slice/crio-008b8ea86c8c2d088af35376717390f5dc36672effe7d166e247b1ad3b0e2e1a WatchSource:0}: Error finding container 008b8ea86c8c2d088af35376717390f5dc36672effe7d166e247b1ad3b0e2e1a: Status 404 returned error can't find the container with id 008b8ea86c8c2d088af35376717390f5dc36672effe7d166e247b1ad3b0e2e1a Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.367238 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f9lb8"] Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.367961 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zbl4m"] Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.368139 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.368708 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.370355 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.373132 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.392875 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.393057 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.393568 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.393793 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.394584 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.395057 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.395537 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.410540 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.425114 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.444176 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.455998 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.468859 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5bd40cdf864fd58bb47516df73d27a7374484fc5768243472656ff258eba40bb"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.470973 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471008 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d903149f5d6d090f01a1d9c061096907e90524df7d4301044df7e0409924b732"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471142 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471190 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-systemd\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471222 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-ovn-kubernetes\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-var-lib-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471291 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-etc-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471348 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-config\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471376 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-os-release\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471732 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-netns\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-ovn\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-node-log\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471852 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-cnibin\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471869 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-log-socket\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471889 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-system-cni-dir\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471931 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e091b0f3-e906-47c0-93a5-5df598bc954e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471948 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-env-overrides\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471963 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-slash\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471979 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-netd\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.471996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjzj5\" (UniqueName: \"kubernetes.io/projected/e091b0f3-e906-47c0-93a5-5df598bc954e-kube-api-access-bjzj5\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472015 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-bin\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472052 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-script-lib\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e091b0f3-e906-47c0-93a5-5df598bc954e-cni-binary-copy\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-systemd-units\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovn-node-metrics-cert\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472399 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-kubelet\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.472485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4g4l\" (UniqueName: \"kubernetes.io/projected/53d5d732-edab-4a9d-802c-4d9f0a2197ac-kube-api-access-c4g4l\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.475612 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.478346 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-t5r7v" event={"ID":"5458a69d-5519-4863-baf9-e549e5e52b78","Type":"ContainerStarted","Data":"0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.478445 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-t5r7v" event={"ID":"5458a69d-5519-4863-baf9-e549e5e52b78","Type":"ContainerStarted","Data":"5586e812952730e5a03720a1d3f5f69a385d204d43fc4b31a97c18c9f37f9e0e"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.480603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.480814 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.480906 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"25e739cdc7413b891671f4acb013457a311a7b85123aa8c53caae7aacf7fadf0"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.482586 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.494094 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.494420 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.499000 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.499538 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.499575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"805fce551e99f5a848a4790d67560bcae2e165aa20870589212857437e590093"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.500947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerStarted","Data":"43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.500974 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerStarted","Data":"008b8ea86c8c2d088af35376717390f5dc36672effe7d166e247b1ad3b0e2e1a"} Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.516622 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.539184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.562746 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.572922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-config\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.572965 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-var-lib-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.572984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-etc-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573029 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-os-release\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573052 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-netns\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573077 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-ovn\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-node-log\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573115 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-cnibin\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573132 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-log-socket\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573146 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-env-overrides\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-system-cni-dir\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573195 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e091b0f3-e906-47c0-93a5-5df598bc954e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573212 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-slash\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573238 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-netd\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573284 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjzj5\" (UniqueName: \"kubernetes.io/projected/e091b0f3-e906-47c0-93a5-5df598bc954e-kube-api-access-bjzj5\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573308 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-bin\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-script-lib\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e091b0f3-e906-47c0-93a5-5df598bc954e-cni-binary-copy\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573376 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-systemd-units\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573408 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovn-node-metrics-cert\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573423 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-kubelet\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573439 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4g4l\" (UniqueName: \"kubernetes.io/projected/53d5d732-edab-4a9d-802c-4d9f0a2197ac-kube-api-access-c4g4l\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573502 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-systemd\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-ovn-kubernetes\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.573579 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-ovn-kubernetes\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574170 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-ovn\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574226 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-bin\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-var-lib-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574190 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-systemd-units\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574287 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-etc-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574313 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-openvswitch\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-slash\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-netd\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574624 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-os-release\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574843 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-node-log\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574913 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-script-lib\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.574951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-netns\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575152 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575269 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-kubelet\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575309 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575338 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-systemd\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-cnibin\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575392 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-log-socket\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575412 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e091b0f3-e906-47c0-93a5-5df598bc954e-system-cni-dir\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575461 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e091b0f3-e906-47c0-93a5-5df598bc954e-cni-binary-copy\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-env-overrides\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.575828 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-config\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.576036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e091b0f3-e906-47c0-93a5-5df598bc954e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.577323 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovn-node-metrics-cert\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.584436 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.599309 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjzj5\" (UniqueName: \"kubernetes.io/projected/e091b0f3-e906-47c0-93a5-5df598bc954e-kube-api-access-bjzj5\") pod \"multus-additional-cni-plugins-zbl4m\" (UID: \"e091b0f3-e906-47c0-93a5-5df598bc954e\") " pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.602895 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4g4l\" (UniqueName: \"kubernetes.io/projected/53d5d732-edab-4a9d-802c-4d9f0a2197ac-kube-api-access-c4g4l\") pod \"ovnkube-node-f9lb8\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.603168 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.618987 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.632018 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.659539 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.700590 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.725948 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.737636 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.740215 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: W1126 06:25:38.740329 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53d5d732_edab_4a9d_802c_4d9f0a2197ac.slice/crio-1a2e13650cc86235ac977839ba308136ea2049d89bc454f98bd3cf310e6af2b4 WatchSource:0}: Error finding container 1a2e13650cc86235ac977839ba308136ea2049d89bc454f98bd3cf310e6af2b4: Status 404 returned error can't find the container with id 1a2e13650cc86235ac977839ba308136ea2049d89bc454f98bd3cf310e6af2b4 Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.778640 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.822182 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.868985 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.875700 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.875813 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:25:40.875792766 +0000 UTC m=+24.237096718 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.875866 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.875932 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.876039 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.876079 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:40.876068594 +0000 UTC m=+24.237372546 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.876187 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.876232 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:40.876223728 +0000 UTC m=+24.237527680 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.901087 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.945135 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.976998 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.977042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977126 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977128 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977159 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977179 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977241 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:40.977221597 +0000 UTC m=+24.338525559 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977141 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977700 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:38 crc kubenswrapper[4775]: E1126 06:25:38.977762 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:40.977750721 +0000 UTC m=+24.339054693 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:38 crc kubenswrapper[4775]: I1126 06:25:38.983011 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:38Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.026384 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.066276 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.119513 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.327135 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:39 crc kubenswrapper[4775]: E1126 06:25:39.327249 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.327155 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:39 crc kubenswrapper[4775]: E1126 06:25:39.327591 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.327677 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:39 crc kubenswrapper[4775]: E1126 06:25:39.327749 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.330884 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.331550 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.332672 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.333418 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.334597 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.335270 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.335898 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.336881 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.337696 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.338619 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.339232 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.340297 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.340885 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.341368 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.342300 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.342801 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.343948 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.344378 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.344986 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.346212 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.346689 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.347635 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.348140 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.349179 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.349623 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.350202 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.351302 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.351844 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.352997 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.353554 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.355300 4775 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.355432 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.357307 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.358265 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.358700 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.360501 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.361217 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.362116 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.362748 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.363772 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.364323 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.365339 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.365965 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.366927 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.367370 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.368291 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.368906 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.370161 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.370696 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.371541 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.372013 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.372910 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.373446 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.374027 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.503919 4775 generic.go:334] "Generic (PLEG): container finished" podID="e091b0f3-e906-47c0-93a5-5df598bc954e" containerID="b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5" exitCode=0 Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.503986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerDied","Data":"b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5"} Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.504014 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerStarted","Data":"b631dbcfe29bdabffbcb273831374bd1b93b5d83f1dc8206ada46d10d884ddd5"} Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.506477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b"} Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.508200 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" exitCode=0 Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.508281 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.508323 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"1a2e13650cc86235ac977839ba308136ea2049d89bc454f98bd3cf310e6af2b4"} Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.531884 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.563798 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.577795 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.599537 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.633083 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.648746 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.664401 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.677638 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.690239 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.702830 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.716292 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.730683 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.747325 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.776496 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.790087 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.802519 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.816963 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.831799 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.856968 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.905215 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.947660 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.964055 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-456fr"] Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.964501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.988853 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fa50ee33-2e13-4179-bfc4-37cad702ef1b-serviceca\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.988938 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdtfx\" (UniqueName: \"kubernetes.io/projected/fa50ee33-2e13-4179-bfc4-37cad702ef1b-kube-api-access-zdtfx\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.988985 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa50ee33-2e13-4179-bfc4-37cad702ef1b-host\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.990676 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:39Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:39 crc kubenswrapper[4775]: I1126 06:25:39.990736 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.010348 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.031536 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.050108 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.089328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fa50ee33-2e13-4179-bfc4-37cad702ef1b-serviceca\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.089368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdtfx\" (UniqueName: \"kubernetes.io/projected/fa50ee33-2e13-4179-bfc4-37cad702ef1b-kube-api-access-zdtfx\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.089398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa50ee33-2e13-4179-bfc4-37cad702ef1b-host\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.089450 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa50ee33-2e13-4179-bfc4-37cad702ef1b-host\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.090436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/fa50ee33-2e13-4179-bfc4-37cad702ef1b-serviceca\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.104823 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.139929 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdtfx\" (UniqueName: \"kubernetes.io/projected/fa50ee33-2e13-4179-bfc4-37cad702ef1b-kube-api-access-zdtfx\") pod \"node-ca-456fr\" (UID: \"fa50ee33-2e13-4179-bfc4-37cad702ef1b\") " pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.162184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.200833 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.240333 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.282464 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-456fr" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.285090 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: W1126 06:25:40.303202 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa50ee33_2e13_4179_bfc4_37cad702ef1b.slice/crio-fffdf9eeb09a3ea818df48abf32b7b2bcee4b88d42b9bc074c23721c4564c6b1 WatchSource:0}: Error finding container fffdf9eeb09a3ea818df48abf32b7b2bcee4b88d42b9bc074c23721c4564c6b1: Status 404 returned error can't find the container with id fffdf9eeb09a3ea818df48abf32b7b2bcee4b88d42b9bc074c23721c4564c6b1 Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.318215 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.362800 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.412583 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.441983 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.484579 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.514798 4775 generic.go:334] "Generic (PLEG): container finished" podID="e091b0f3-e906-47c0-93a5-5df598bc954e" containerID="147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a" exitCode=0 Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.514879 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerDied","Data":"147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.521017 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.523325 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.523370 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.523381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.523394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.523405 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.525158 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-456fr" event={"ID":"fa50ee33-2e13-4179-bfc4-37cad702ef1b","Type":"ContainerStarted","Data":"fffdf9eeb09a3ea818df48abf32b7b2bcee4b88d42b9bc074c23721c4564c6b1"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.527130 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203"} Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.564554 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.602443 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.637667 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.681100 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.728235 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.761877 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.800970 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.845206 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.880630 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.900085 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.900238 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:40 crc kubenswrapper[4775]: E1126 06:25:40.900382 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:25:44.900338413 +0000 UTC m=+28.261642405 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:25:40 crc kubenswrapper[4775]: E1126 06:25:40.900457 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.900583 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:40 crc kubenswrapper[4775]: E1126 06:25:40.900617 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:44.90058058 +0000 UTC m=+28.261884672 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:40 crc kubenswrapper[4775]: E1126 06:25:40.900835 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:40 crc kubenswrapper[4775]: E1126 06:25:40.900963 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:44.900938569 +0000 UTC m=+28.262242701 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.920456 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:40 crc kubenswrapper[4775]: I1126 06:25:40.964281 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:40Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.001298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.001403 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001570 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001568 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001596 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001620 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001627 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001643 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001708 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:45.001686122 +0000 UTC m=+28.362990114 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.001766 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:45.001752933 +0000 UTC m=+28.363056915 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.006221 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.044925 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.084940 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.126592 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.167416 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.203505 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.246912 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.282015 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.321157 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.327470 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.327600 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.327636 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.327842 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.327996 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:41 crc kubenswrapper[4775]: E1126 06:25:41.328256 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.365126 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.537086 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.539925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-456fr" event={"ID":"fa50ee33-2e13-4179-bfc4-37cad702ef1b","Type":"ContainerStarted","Data":"d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e"} Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.544623 4775 generic.go:334] "Generic (PLEG): container finished" podID="e091b0f3-e906-47c0-93a5-5df598bc954e" containerID="ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446" exitCode=0 Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.544761 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerDied","Data":"ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446"} Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.568183 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.585110 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.600741 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.623603 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.645177 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.666157 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.686675 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.719151 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.745757 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.760464 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.808955 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.847035 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.886003 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.920641 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:41 crc kubenswrapper[4775]: I1126 06:25:41.961972 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.000265 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:41Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.042210 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.079428 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.118223 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.162483 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.201437 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.238429 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.278570 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.337545 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.363589 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.401012 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.440773 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.483114 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.553153 4775 generic.go:334] "Generic (PLEG): container finished" podID="e091b0f3-e906-47c0-93a5-5df598bc954e" containerID="2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e" exitCode=0 Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.553381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerDied","Data":"2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e"} Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.577501 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.604492 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.624868 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.646307 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.685863 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.723557 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.772585 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.806802 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.839759 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.888924 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.926812 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.959598 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:42 crc kubenswrapper[4775]: I1126 06:25:42.999123 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:42Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.024003 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.026415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.026454 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.026464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.026519 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.043271 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.072424 4775 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.073046 4775 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.074503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.074534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.074547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.074566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.074579 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.091929 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.096609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.096637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.096646 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.096660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.096670 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.116437 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.120541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.120584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.120595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.120617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.120631 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.142620 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.146796 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.146838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.146847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.146865 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.146876 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.165546 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.171149 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.171204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.171217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.171242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.171256 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.190138 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.190441 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.192945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.193013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.193028 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.193054 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.193072 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.297105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.297675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.297694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.297744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.297760 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.327275 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.327424 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.327476 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.327551 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.327672 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:43 crc kubenswrapper[4775]: E1126 06:25:43.327802 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.402231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.402768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.402975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.403166 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.406574 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.509924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.510016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.510036 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.510066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.510086 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.566035 4775 generic.go:334] "Generic (PLEG): container finished" podID="e091b0f3-e906-47c0-93a5-5df598bc954e" containerID="bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173" exitCode=0 Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.566108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerDied","Data":"bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.576643 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.588778 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.613017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.613041 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.613100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.613316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.613344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.613362 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.638650 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.663817 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.701799 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.717941 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.718938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.718980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.718998 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.719028 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.719048 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.736775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.748846 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.761052 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.809541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.821802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.821849 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.821862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.821882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.821897 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.845415 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.858347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.878246 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.895986 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:43Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.925592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.925627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.925637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.925651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:43 crc kubenswrapper[4775]: I1126 06:25:43.925660 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:43Z","lastTransitionTime":"2025-11-26T06:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.028388 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.028422 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.028433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.028448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.028459 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.130775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.130810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.130822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.130837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.130846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.235293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.235367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.235396 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.235449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.235473 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.337642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.337707 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.337739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.337758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.337771 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.440978 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.441056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.441081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.441112 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.441134 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.543395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.543437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.543452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.543468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.543478 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.582904 4775 generic.go:334] "Generic (PLEG): container finished" podID="e091b0f3-e906-47c0-93a5-5df598bc954e" containerID="85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135" exitCode=0 Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.582961 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerDied","Data":"85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.602240 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.615304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.645646 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.646459 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.646563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.646583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.646608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.646626 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.665302 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.674914 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.694234 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.711012 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.724494 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.738959 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.749635 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.749681 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.749701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.749746 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.749762 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.760399 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.773409 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.786520 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.799382 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.811477 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:44Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.852800 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.852862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.852878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.852899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.852912 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.953842 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954005 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954067 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:44 crc kubenswrapper[4775]: E1126 06:25:44.954190 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:25:52.954143457 +0000 UTC m=+36.315447469 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:25:44 crc kubenswrapper[4775]: E1126 06:25:44.954207 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:44 crc kubenswrapper[4775]: E1126 06:25:44.954235 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:44 crc kubenswrapper[4775]: E1126 06:25:44.954303 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:52.954289541 +0000 UTC m=+36.315593493 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:44 crc kubenswrapper[4775]: E1126 06:25:44.954670 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:52.95463969 +0000 UTC m=+36.315943732 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954905 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954955 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:44 crc kubenswrapper[4775]: I1126 06:25:44.954968 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:44Z","lastTransitionTime":"2025-11-26T06:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.055131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.055208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055322 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055337 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055348 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055397 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:53.055385553 +0000 UTC m=+36.416689505 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055442 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055487 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055509 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.055589 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:53.055564527 +0000 UTC m=+36.416868519 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.057982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.058042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.058058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.058081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.058097 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.160601 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.160653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.160667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.160685 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.160698 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.264179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.264237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.264254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.264278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.264295 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.327632 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.327684 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.327794 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.327824 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.327994 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:45 crc kubenswrapper[4775]: E1126 06:25:45.328079 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.367509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.367565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.367577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.367595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.367610 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.470907 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.470960 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.470973 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.470991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.471005 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.574049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.574095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.574107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.574126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.574139 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.591584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.592041 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.592087 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.599178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" event={"ID":"e091b0f3-e906-47c0-93a5-5df598bc954e","Type":"ContainerStarted","Data":"ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.607968 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.632793 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.633668 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.636787 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.650904 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.671527 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.677498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.677559 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.677578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.677607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.677625 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.685903 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.717752 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.743147 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.759163 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.780083 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.780830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.781079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.781104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.781129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.781147 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.800676 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.821420 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.842918 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.861462 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.876903 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.884100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.884148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.884166 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.884189 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.884206 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.891558 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.915012 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.936529 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.959900 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.980014 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.987379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.987444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.987463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.987490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.987510 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:45Z","lastTransitionTime":"2025-11-26T06:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:45 crc kubenswrapper[4775]: I1126 06:25:45.997382 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:45Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.019759 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.042949 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.062566 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.084466 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.090857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.090939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.090951 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.090989 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.091003 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.120748 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.148063 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.163501 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.181691 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:46Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.194091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.194146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.194160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.194181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.194199 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.297634 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.297680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.297697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.297753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.297772 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.401798 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.401870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.401891 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.401916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.401933 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.505884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.505947 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.505964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.505991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.506008 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.603496 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.608941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.609017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.609042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.609073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.609097 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.713401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.713503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.713529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.713602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.713627 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.818004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.818125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.818144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.818170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.818187 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.921744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.921803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.921821 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.921848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:46 crc kubenswrapper[4775]: I1126 06:25:46.921864 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:46Z","lastTransitionTime":"2025-11-26T06:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.025078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.025147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.025159 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.025181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.025193 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.128631 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.128701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.128757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.128783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.128801 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.230804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.230885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.230901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.230924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.230940 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.327297 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.327371 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.327493 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:47 crc kubenswrapper[4775]: E1126 06:25:47.327581 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:47 crc kubenswrapper[4775]: E1126 06:25:47.327689 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:47 crc kubenswrapper[4775]: E1126 06:25:47.327809 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.332965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.333036 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.333060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.333092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.333114 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.346876 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.367369 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.387021 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.406115 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.418523 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.432679 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.434824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.434856 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.434868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.434886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.434899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.444854 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.455359 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.464540 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.475844 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.484208 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.499819 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.513481 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.530477 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:47Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.537192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.537226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.537235 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.537255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.537267 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.606902 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.641085 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.641374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.641391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.641414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.641430 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.743999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.744071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.744089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.744118 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.744136 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.846474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.846539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.846556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.846580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.846598 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.949439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.949498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.949515 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.949541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:47 crc kubenswrapper[4775]: I1126 06:25:47.949559 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:47Z","lastTransitionTime":"2025-11-26T06:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.052209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.052262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.052279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.052302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.052320 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.155304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.155366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.155384 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.155409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.155426 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.258469 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.258520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.258537 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.258560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.258577 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.361395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.361452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.361468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.361492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.361509 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.464580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.464643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.464662 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.465135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.465488 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.568503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.568547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.568557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.568576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.568587 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.613077 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/0.log" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.617144 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229" exitCode=1 Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.617209 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.618375 4775 scope.go:117] "RemoveContainer" containerID="af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.645867 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.672525 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.672624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.672690 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.672749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.672785 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.672810 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.693462 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.712521 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.728058 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.748483 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.769753 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.775680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.775763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.775781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.775809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.775830 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.787922 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.809606 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.840501 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.868538 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.878521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.878578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.878593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.878611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.878623 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.889553 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.908127 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.924770 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:48Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.981000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.981073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.981096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.981125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:48 crc kubenswrapper[4775]: I1126 06:25:48.981148 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:48Z","lastTransitionTime":"2025-11-26T06:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.084089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.084147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.084169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.084195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.084216 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.186707 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.186792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.186808 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.186833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.186855 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.289621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.289679 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.289697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.289749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.289772 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.327390 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.327494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:49 crc kubenswrapper[4775]: E1126 06:25:49.327845 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.327878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:49 crc kubenswrapper[4775]: E1126 06:25:49.327967 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:49 crc kubenswrapper[4775]: E1126 06:25:49.328068 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.391827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.391869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.391880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.391898 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.391909 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.495202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.495258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.495271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.495291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.495305 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.598249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.598309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.598317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.598338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.598352 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.623893 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/0.log" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.627233 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.627392 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.646766 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.664131 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.685274 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.700944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.700991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.701004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.701023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.701037 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.701258 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.716789 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.738486 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.756374 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.774777 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.784474 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.799551 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.804162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.804204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.804217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.804234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.804249 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.816886 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.837334 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.857310 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.871849 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:49Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.907539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.907609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.907627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.907653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:49 crc kubenswrapper[4775]: I1126 06:25:49.907673 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:49Z","lastTransitionTime":"2025-11-26T06:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.010973 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.011051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.011073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.011105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.011126 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.059555 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.074537 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.093438 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.112983 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.113914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.113964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.113978 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.114001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.114014 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.134692 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.163783 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.187710 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.208240 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.217126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.217169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.217180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.217199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.217212 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.231823 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.250964 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.273070 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.293330 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.313757 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.319886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.319952 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.319974 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.320002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.320021 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.333643 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.350198 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.423524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.423604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.423627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.423661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.423686 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.528384 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.528450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.528468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.528491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.528508 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.631254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.631330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.631357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.632002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.632043 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.634439 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/1.log" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.635598 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/0.log" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.641134 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef" exitCode=1 Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.641229 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.641318 4775 scope.go:117] "RemoveContainer" containerID="af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.642580 4775 scope.go:117] "RemoveContainer" containerID="c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef" Nov 26 06:25:50 crc kubenswrapper[4775]: E1126 06:25:50.642929 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.668073 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.690685 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.695605 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx"] Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.700299 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.708055 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.708070 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.711428 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.725608 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.735516 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.735554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.735572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.735598 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.735619 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.746249 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.766129 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.785659 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.803841 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.827570 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5739dc36-9102-437f-95c0-b9e137f4a458-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.827632 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hcx9\" (UniqueName: \"kubernetes.io/projected/5739dc36-9102-437f-95c0-b9e137f4a458-kube-api-access-2hcx9\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.827662 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5739dc36-9102-437f-95c0-b9e137f4a458-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.827683 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5739dc36-9102-437f-95c0-b9e137f4a458-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.833161 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.838444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.838503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.838522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.838547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.838566 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.855496 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.872768 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.892999 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.912448 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.928228 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5739dc36-9102-437f-95c0-b9e137f4a458-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.928291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5739dc36-9102-437f-95c0-b9e137f4a458-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.928371 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5739dc36-9102-437f-95c0-b9e137f4a458-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.928452 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hcx9\" (UniqueName: \"kubernetes.io/projected/5739dc36-9102-437f-95c0-b9e137f4a458-kube-api-access-2hcx9\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.929265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5739dc36-9102-437f-95c0-b9e137f4a458-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.929326 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5739dc36-9102-437f-95c0-b9e137f4a458-env-overrides\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.931017 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.937109 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5739dc36-9102-437f-95c0-b9e137f4a458-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.941655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.941692 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.941703 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.941747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.941764 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:50Z","lastTransitionTime":"2025-11-26T06:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.947138 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hcx9\" (UniqueName: \"kubernetes.io/projected/5739dc36-9102-437f-95c0-b9e137f4a458-kube-api-access-2hcx9\") pod \"ovnkube-control-plane-749d76644c-tmqbx\" (UID: \"5739dc36-9102-437f-95c0-b9e137f4a458\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.950649 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.969347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.985847 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:50 crc kubenswrapper[4775]: I1126 06:25:50.999832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:50Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.011601 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.022986 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.029340 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.045053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.045647 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.045664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.045681 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.045693 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.046536 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.063383 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.088406 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.112709 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.129542 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.154191 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.154252 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.154271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.154299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.154317 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.173347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.210429 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.230574 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.255814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.255847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.255857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.255872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.255883 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.261321 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.326906 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.326910 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.327020 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:51 crc kubenswrapper[4775]: E1126 06:25:51.327157 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:51 crc kubenswrapper[4775]: E1126 06:25:51.327280 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:51 crc kubenswrapper[4775]: E1126 06:25:51.327464 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.364527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.364597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.364617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.364685 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.364703 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.469262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.469324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.469342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.469368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.469386 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.572288 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.572349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.572366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.572392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.572408 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.655519 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/1.log" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.661389 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" event={"ID":"5739dc36-9102-437f-95c0-b9e137f4a458","Type":"ContainerStarted","Data":"dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.661472 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" event={"ID":"5739dc36-9102-437f-95c0-b9e137f4a458","Type":"ContainerStarted","Data":"4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.661494 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" event={"ID":"5739dc36-9102-437f-95c0-b9e137f4a458","Type":"ContainerStarted","Data":"e654d38b4c4d889ba92ce6e060d6a45871edcf1b5adb06b5c7b4bec42b9b072c"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.674424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.674467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.674478 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.674495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.674506 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.677867 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.692858 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.710347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.725783 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.738105 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.756820 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.768429 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.776429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.776589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.776697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.776833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.776949 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.785061 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.797520 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.815050 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.840026 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.861664 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.875495 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.880173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.880210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.880226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.880249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.880267 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.892078 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.907118 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:51Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.982795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.983153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.983302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.983453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:51 crc kubenswrapper[4775]: I1126 06:25:51.983617 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:51Z","lastTransitionTime":"2025-11-26T06:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.085985 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.086046 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.086058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.086077 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.086095 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.189769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.189829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.189845 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.189870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.189886 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.245628 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-n98r8"] Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.246444 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: E1126 06:25:52.246557 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.264450 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.285184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.292751 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.292806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.292817 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.292838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.292852 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.306969 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.321908 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.335299 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.344158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.344272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz7rv\" (UniqueName: \"kubernetes.io/projected/1e24b0f5-8555-40a4-a965-e67d5cd305a6-kube-api-access-jz7rv\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.346267 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.357847 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.368272 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.386892 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.394954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.395010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.395047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.395065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.395079 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.406184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.421042 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.439910 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af8214f498c46465b15bd50fe1cb4c3119d01a90f352f4ed1a08dcefdaee4229\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:47Z\\\",\\\"message\\\":\\\"o:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 06:25:47.788948 6084 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 06:25:47.788977 6084 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 06:25:47.789036 6084 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 06:25:47.789058 6084 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1126 06:25:47.789086 6084 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1126 06:25:47.789093 6084 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 06:25:47.789112 6084 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 06:25:47.789121 6084 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 06:25:47.789143 6084 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 06:25:47.789159 6084 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 06:25:47.789163 6084 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 06:25:47.789179 6084 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 06:25:47.789182 6084 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1126 06:25:47.789179 6084 factory.go:656] Stopping watch factory\\\\nI1126 06:25:47.789193 6084 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.445216 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz7rv\" (UniqueName: \"kubernetes.io/projected/1e24b0f5-8555-40a4-a965-e67d5cd305a6-kube-api-access-jz7rv\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.445281 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: E1126 06:25:52.445440 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:52 crc kubenswrapper[4775]: E1126 06:25:52.445527 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:52.945504944 +0000 UTC m=+36.306808886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.457296 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.469519 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz7rv\" (UniqueName: \"kubernetes.io/projected/1e24b0f5-8555-40a4-a965-e67d5cd305a6-kube-api-access-jz7rv\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.474737 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.489039 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.497794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.497842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.497852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.497872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.497884 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.503961 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:52Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.601024 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.601083 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.601100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.601126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.601147 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.704195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.704267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.704283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.704307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.704322 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.807837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.807913 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.807933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.807963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.807981 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.910696 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.910773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.910782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.910801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.910812 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:52Z","lastTransitionTime":"2025-11-26T06:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:52 crc kubenswrapper[4775]: I1126 06:25:52.952403 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:52 crc kubenswrapper[4775]: E1126 06:25:52.952619 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:52 crc kubenswrapper[4775]: E1126 06:25:52.952708 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:53.952681437 +0000 UTC m=+37.313985589 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.013934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.014005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.014016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.014039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.014057 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.052940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.053141 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.053198 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.053286 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:26:09.053230424 +0000 UTC m=+52.414534416 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.053419 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.053551 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:09.053521092 +0000 UTC m=+52.414825084 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.053981 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.054094 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:09.054071366 +0000 UTC m=+52.415375358 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.117200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.117267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.117285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.117311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.117331 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.154777 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.154879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155093 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155120 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155141 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155207 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:09.155185939 +0000 UTC m=+52.516489921 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155242 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155289 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155310 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.155390 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:09.155363373 +0000 UTC m=+52.516667355 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.192399 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.192465 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.192487 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.192513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.192534 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.219269 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:53Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.225670 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.225767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.225792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.225823 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.225846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.247907 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:53Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.253989 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.254105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.254127 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.254154 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.254193 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.272396 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:53Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.278416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.278501 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.278523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.278557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.278583 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.300506 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:53Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.308120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.308186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.308207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.308234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.308255 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.327992 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.328097 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.328114 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.328276 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.328366 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.328483 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.328593 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.328689 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.333513 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:53Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.333897 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.336993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.337058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.337080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.337105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.337126 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.452132 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.452199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.452218 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.452252 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.452275 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.555690 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.555793 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.555809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.555832 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.555846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.659157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.659229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.659238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.659257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.659271 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.762231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.762331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.762351 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.762387 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.762411 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.866156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.866292 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.866322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.866353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.866380 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.965180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.965492 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: E1126 06:25:53.965684 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:55.965644448 +0000 UTC m=+39.326948440 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.970175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.970355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.970376 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.970401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:53 crc kubenswrapper[4775]: I1126 06:25:53.970423 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:53Z","lastTransitionTime":"2025-11-26T06:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.074948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.075020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.075041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.075066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.075083 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.178975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.179047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.179064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.179093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.179110 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.282158 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.282238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.282255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.282282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.282301 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.385702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.385778 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.385791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.385811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.385826 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.488952 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.489001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.489017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.489040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.489056 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.592220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.592297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.592322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.592358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.592384 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.695607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.695674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.695695 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.695759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.695786 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.799250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.799357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.799376 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.799400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.799420 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.902982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.903077 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.903106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.903137 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:54 crc kubenswrapper[4775]: I1126 06:25:54.903156 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:54Z","lastTransitionTime":"2025-11-26T06:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.006461 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.006553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.006578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.006604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.006621 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.110203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.110273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.110290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.110314 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.110330 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.213072 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.213141 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.213160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.213188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.213210 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.316964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.317021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.317033 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.317054 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.317072 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.327625 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.327745 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.327797 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.328026 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:55 crc kubenswrapper[4775]: E1126 06:25:55.328006 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:25:55 crc kubenswrapper[4775]: E1126 06:25:55.328299 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:55 crc kubenswrapper[4775]: E1126 06:25:55.328434 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:55 crc kubenswrapper[4775]: E1126 06:25:55.328641 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.420509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.420580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.420598 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.420649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.420673 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.524592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.525213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.525345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.525418 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.525477 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.629143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.629568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.629649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.629774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.629846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.732263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.732313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.732326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.732345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.732356 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.835135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.835178 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.835209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.835243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.835257 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.939230 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.939334 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.939355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.939381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.939399 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:55Z","lastTransitionTime":"2025-11-26T06:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:55 crc kubenswrapper[4775]: E1126 06:25:55.991229 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:55 crc kubenswrapper[4775]: E1126 06:25:55.991353 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:25:59.991320803 +0000 UTC m=+43.352624795 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:25:55 crc kubenswrapper[4775]: I1126 06:25:55.991062 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.042440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.042489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.042505 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.042527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.042544 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.147063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.147120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.147136 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.147160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.147182 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.250307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.250371 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.250385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.250410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.250426 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.353668 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.353805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.353826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.353852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.353872 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.457780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.457843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.457860 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.457885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.457903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.477015 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.478548 4775 scope.go:117] "RemoveContainer" containerID="c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef" Nov 26 06:25:56 crc kubenswrapper[4775]: E1126 06:25:56.479030 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.510394 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.534434 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.550412 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.560671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.561021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.561285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.561453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.561575 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.570125 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.589744 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.606252 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.627035 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.650136 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.665014 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.665087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.665106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.665134 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.665152 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.677576 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.695479 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.717681 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.735305 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.757813 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.767876 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.768230 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.768388 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.768527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.768646 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.776668 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.800011 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.820008 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:56Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.871813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.871866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.871883 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.871912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.871931 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.975200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.975358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.975391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.975442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:56 crc kubenswrapper[4775]: I1126 06:25:56.975466 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:56Z","lastTransitionTime":"2025-11-26T06:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.078190 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.078656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.078929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.079095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.079223 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.181863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.181926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.181948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.181972 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.181989 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.285684 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.285767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.285781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.285804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.285819 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.327406 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.327506 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.327544 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:57 crc kubenswrapper[4775]: E1126 06:25:57.327682 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:57 crc kubenswrapper[4775]: E1126 06:25:57.327910 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.327934 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:57 crc kubenswrapper[4775]: E1126 06:25:57.327986 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:25:57 crc kubenswrapper[4775]: E1126 06:25:57.328089 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.354677 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.378572 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.389073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.389137 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.389155 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.389180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.389200 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.396530 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.416322 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.438711 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.462334 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.483964 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.492337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.492416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.492442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.492474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.492501 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.506031 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.524222 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.541911 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.561224 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.577904 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.595207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.595262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.595284 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.595315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.595335 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.596633 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.618342 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.639754 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.657246 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:25:57Z is after 2025-08-24T17:21:41Z" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.698204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.698267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.698284 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.698309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.698327 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.801462 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.801530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.801548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.801575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.801594 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.905520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.905632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.905651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.905676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:57 crc kubenswrapper[4775]: I1126 06:25:57.905695 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:57Z","lastTransitionTime":"2025-11-26T06:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.009653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.009746 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.009767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.009797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.009816 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.133759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.133818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.133838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.133863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.133884 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.237324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.237383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.237400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.237426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.237445 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.340766 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.340824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.340840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.340869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.340891 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.444486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.444551 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.444568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.444594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.444612 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.548181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.548252 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.548274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.548304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.548326 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.651882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.651958 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.651975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.652001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.652020 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.756242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.756298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.756314 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.756339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.756355 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.859005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.859082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.859105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.859137 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.859159 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.962397 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.962505 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.962530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.962555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:58 crc kubenswrapper[4775]: I1126 06:25:58.962574 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:58Z","lastTransitionTime":"2025-11-26T06:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.065848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.065915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.065932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.065956 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.065974 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.169010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.169068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.169086 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.169109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.169125 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.271585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.271648 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.271665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.271690 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.271707 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.326634 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.326778 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:25:59 crc kubenswrapper[4775]: E1126 06:25:59.326849 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.326917 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:25:59 crc kubenswrapper[4775]: E1126 06:25:59.327050 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.327126 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:25:59 crc kubenswrapper[4775]: E1126 06:25:59.327321 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:25:59 crc kubenswrapper[4775]: E1126 06:25:59.327478 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.374417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.374493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.374520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.374551 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.374574 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.477894 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.477948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.477972 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.477999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.478020 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.580934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.581008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.581048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.581077 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.581093 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.684204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.684243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.684258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.684275 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.684287 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.786795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.786859 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.786875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.786899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.786950 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.889920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.889981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.889999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.890021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.890041 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.993346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.993420 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.993439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.993464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:25:59 crc kubenswrapper[4775]: I1126 06:25:59.993482 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:25:59Z","lastTransitionTime":"2025-11-26T06:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.056679 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:00 crc kubenswrapper[4775]: E1126 06:26:00.057024 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:00 crc kubenswrapper[4775]: E1126 06:26:00.057109 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:08.057088135 +0000 UTC m=+51.418392127 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.096280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.096340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.096364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.096392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.096414 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.199684 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.199767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.199785 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.199810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.199836 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.304290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.304338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.304353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.304377 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.304394 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.407261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.407320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.407338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.407363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.407380 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.510164 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.510237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.510254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.510281 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.510299 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.613983 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.614047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.614065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.614092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.614118 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.716657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.716755 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.716776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.716802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.716846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.819414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.819499 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.819521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.819554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.819577 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.922567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.922868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.922893 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.922916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:00 crc kubenswrapper[4775]: I1126 06:26:00.922936 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:00Z","lastTransitionTime":"2025-11-26T06:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.026063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.026103 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.026120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.026142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.026158 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.129052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.129109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.129133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.129160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.129181 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.231970 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.232031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.232052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.232081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.232104 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.327268 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.327370 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:01 crc kubenswrapper[4775]: E1126 06:26:01.327831 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.327556 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:01 crc kubenswrapper[4775]: E1126 06:26:01.328339 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:01 crc kubenswrapper[4775]: E1126 06:26:01.328041 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.327384 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:01 crc kubenswrapper[4775]: E1126 06:26:01.328868 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.335332 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.335409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.335427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.335487 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.335506 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.438358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.438416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.438434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.438458 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.438475 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.541560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.541624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.541643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.541670 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.541688 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.645258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.645323 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.645344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.645372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.645390 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.748830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.748906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.748927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.748959 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.748981 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.852201 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.852279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.852302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.852334 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.852356 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.956202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.956242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.956258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.956275 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:01 crc kubenswrapper[4775]: I1126 06:26:01.956289 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:01Z","lastTransitionTime":"2025-11-26T06:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.059213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.059255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.059267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.059284 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.059297 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.162548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.162618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.162640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.162671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.162695 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.266812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.267201 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.267382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.267510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.267648 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.370597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.370933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.371062 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.371184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.371352 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.474471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.474531 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.474548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.474631 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.474653 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.577330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.577370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.577377 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.577393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.577404 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.681078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.681121 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.681130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.681144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.681157 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.784691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.784756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.784769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.784787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.784800 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.888031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.888093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.888110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.888133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.888333 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.990926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.990963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.990974 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.990990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:02 crc kubenswrapper[4775]: I1126 06:26:02.991001 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:02Z","lastTransitionTime":"2025-11-26T06:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.095047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.095523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.095668 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.095830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.095952 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.199293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.200172 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.200311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.200436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.200567 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.304252 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.304414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.304443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.304472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.304494 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.327054 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.327142 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.327242 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.327500 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.327515 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.327696 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.327933 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.328155 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.407137 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.407203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.407222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.407248 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.407268 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.510710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.510811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.510840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.510873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.510896 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.513920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.514128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.514290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.514443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.514588 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.537015 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:03Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.543218 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.543295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.543317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.543348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.543375 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.566612 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:03Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.572059 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.572321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.572464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.572640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.572810 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.596691 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:03Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.602768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.602825 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.602837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.602857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.602869 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.624736 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:03Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.630020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.630058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.630070 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.630087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.630099 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.648264 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:03Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:03 crc kubenswrapper[4775]: E1126 06:26:03.648405 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.649893 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.649933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.649949 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.649968 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.649983 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.753573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.753636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.753660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.753690 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.753748 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.856653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.856700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.856757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.856798 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.856824 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.960591 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.961018 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.961394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.961648 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:03 crc kubenswrapper[4775]: I1126 06:26:03.961892 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:03Z","lastTransitionTime":"2025-11-26T06:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.065073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.065115 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.065131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.065188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.065205 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.168786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.168872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.168895 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.168925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.168947 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.271514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.271583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.271596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.271612 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.271623 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.374558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.374639 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.374670 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.374707 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.374768 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.481518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.481573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.481589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.481611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.481626 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.584119 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.584197 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.584225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.584447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.584470 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.686748 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.686810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.686827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.686851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.686868 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.790256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.790309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.790327 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.790349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.790366 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.893674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.893743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.893761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.893785 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.893801 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.996764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.996829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.996852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.996882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:04 crc kubenswrapper[4775]: I1126 06:26:04.996903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:04Z","lastTransitionTime":"2025-11-26T06:26:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.099851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.099900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.099917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.099938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.099953 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.202708 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.202791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.202809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.202836 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.202855 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.305065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.305200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.305271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.305301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.305323 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.327231 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.327327 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.327351 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:05 crc kubenswrapper[4775]: E1126 06:26:05.327552 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.327570 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:05 crc kubenswrapper[4775]: E1126 06:26:05.327667 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:05 crc kubenswrapper[4775]: E1126 06:26:05.327895 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:05 crc kubenswrapper[4775]: E1126 06:26:05.328068 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.409224 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.409325 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.409345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.409376 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.409393 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.511861 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.511902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.511916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.511931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.511941 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.615026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.615093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.615115 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.615147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.615168 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.720285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.720352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.720372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.720398 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.720427 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.823879 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.823944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.823963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.823992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.824011 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.927664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.927770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.927799 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.927834 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:05 crc kubenswrapper[4775]: I1126 06:26:05.927857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:05Z","lastTransitionTime":"2025-11-26T06:26:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.030408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.030466 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.030481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.030502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.030516 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.133882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.134002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.134020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.134045 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.134061 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.236641 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.236760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.236785 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.236811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.236829 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.339333 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.339372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.339381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.339393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.339402 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.441270 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.441340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.441363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.441390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.441413 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.543356 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.543401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.543412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.543430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.543447 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.646294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.646335 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.646346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.646363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.646374 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.748768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.748866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.748885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.748944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.748963 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.852519 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.852611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.852631 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.852689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.852761 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.956232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.956302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.956326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.956355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:06 crc kubenswrapper[4775]: I1126 06:26:06.956378 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:06Z","lastTransitionTime":"2025-11-26T06:26:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.059230 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.059288 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.059306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.059330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.059347 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.162650 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.162756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.162812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.162844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.162864 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.266182 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.266244 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.266269 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.266295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.266315 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.327684 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.327819 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:07 crc kubenswrapper[4775]: E1126 06:26:07.327898 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.327916 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:07 crc kubenswrapper[4775]: E1126 06:26:07.328189 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.328293 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:07 crc kubenswrapper[4775]: E1126 06:26:07.328874 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:07 crc kubenswrapper[4775]: E1126 06:26:07.329045 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.329353 4775 scope.go:117] "RemoveContainer" containerID="c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.356239 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.369058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.369104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.369117 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.369138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.369151 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.380442 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.402997 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.428178 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.453036 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.466301 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.471882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.471965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.471986 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.472016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.472038 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.480503 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.497069 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.512033 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.527349 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.547282 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.563673 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.574327 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.574365 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.574383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.574407 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.574425 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.576405 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.587669 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.602928 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.618179 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.676320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.676357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.676365 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.676379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.676389 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.729235 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/1.log" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.736126 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.736752 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.751070 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.766169 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.779090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.779178 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.779195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.779221 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.779239 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.785061 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.805936 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.835391 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.863532 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.881892 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.881942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.881959 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.881981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.881996 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.882047 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.900265 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.913099 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.930766 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.942290 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.954046 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.969858 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.978855 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.983855 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.983903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.983917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.983959 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.983972 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:07Z","lastTransitionTime":"2025-11-26T06:26:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:07 crc kubenswrapper[4775]: I1126 06:26:07.991341 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:07Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.007161 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.086319 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.086581 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.086652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.086737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.086820 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.145396 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:08 crc kubenswrapper[4775]: E1126 06:26:08.145532 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:08 crc kubenswrapper[4775]: E1126 06:26:08.145578 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:24.145563799 +0000 UTC m=+67.506867751 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.189390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.189811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.189995 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.190101 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.190183 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.291795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.292005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.292110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.292178 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.292233 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.394535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.394619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.394642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.394671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.394692 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.498167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.498272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.498291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.498315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.498332 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.601521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.601576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.601592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.601619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.601645 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.704258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.704567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.704694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.704853 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.705025 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.745649 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/2.log" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.748870 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/1.log" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.753837 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297" exitCode=1 Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.754217 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.754308 4775 scope.go:117] "RemoveContainer" containerID="c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.755084 4775 scope.go:117] "RemoveContainer" containerID="023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297" Nov 26 06:26:08 crc kubenswrapper[4775]: E1126 06:26:08.755375 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.780973 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.820783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.820832 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.820847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.820880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.820899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.833805 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.859980 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.873557 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.890159 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.902674 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.916628 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.923213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.923248 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.923262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.923280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.923292 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:08Z","lastTransitionTime":"2025-11-26T06:26:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.930621 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.944671 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.956089 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.967594 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.981140 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:08 crc kubenswrapper[4775]: I1126 06:26:08.992753 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:08Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.007473 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.025666 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.025767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.025786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.025809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.025828 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.027038 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.043925 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.054048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.054157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.054216 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.054297 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:26:41.054279587 +0000 UTC m=+84.415583549 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.054325 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.054398 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:41.05438438 +0000 UTC m=+84.415688352 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.054406 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.054543 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:41.054516423 +0000 UTC m=+84.415820405 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.128607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.128639 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.128660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.128676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.128685 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.155749 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.155815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155918 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155938 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155937 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155949 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155959 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155971 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.155994 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:41.155979205 +0000 UTC m=+84.517283157 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.156013 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:41.156000715 +0000 UTC m=+84.517304677 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.231582 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.231643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.231664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.231693 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.231748 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.327175 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.327263 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.327299 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.327424 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.327437 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.327625 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.327697 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.327801 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.334370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.334420 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.334437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.334462 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.334479 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.437348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.437414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.437432 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.437457 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.437475 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.540594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.540653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.540669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.540695 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.540743 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.643784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.643843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.643860 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.643888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.643906 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.711577 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.728439 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.731946 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.753407 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.754145 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.754313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.754344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.754420 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.754445 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.761106 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/2.log" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.771918 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.772498 4775 scope.go:117] "RemoveContainer" containerID="023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297" Nov 26 06:26:09 crc kubenswrapper[4775]: E1126 06:26:09.772961 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.793665 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.821532 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4375f0f41722672d2e16d4a485c8a1de7e7bafd4d4113f1c040e287086b83ef\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:25:49Z\\\",\\\"message\\\":\\\"{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-ingress-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-ingress-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.244\\\\\\\", Port:9393, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1126 06:25:49.666427 6199 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-ingress-canary/ingress-canary]} name:Service_openshift-ingress-canary/ingress-canary_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.34:8443: 10.217.5.34:8888:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7715118b-bb1b-400a-803e-7ab2cc3eeec0}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.846029 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.860163 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.860453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.860602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.860861 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.861065 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.869184 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.888485 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.904400 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.926359 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.946158 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.963510 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.964226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.964342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.964369 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.964400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.964427 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:09Z","lastTransitionTime":"2025-11-26T06:26:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.978958 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:09 crc kubenswrapper[4775]: I1126 06:26:09.992472 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:09Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.006875 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.021323 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.037201 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.054119 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.067403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.067466 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.067484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.067512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.067535 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.072961 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.090477 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.109083 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.129956 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.151906 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.167329 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.170487 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.170526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.170536 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.170553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.170567 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.180618 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.193843 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.210118 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.225166 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.247421 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.269896 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.273241 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.273345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.273363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.273386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.273403 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.287625 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.302029 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.323314 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:10Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.376148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.376217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.376242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.376274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.376295 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.479453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.479520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.479538 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.479563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.479580 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.582414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.582468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.582486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.582510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.582532 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.686286 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.686342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.686360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.686386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.686404 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.789403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.789464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.789480 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.789503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.789520 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.892858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.892919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.892938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.892962 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.892980 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.996271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.996329 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.996347 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.996371 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:10 crc kubenswrapper[4775]: I1126 06:26:10.996389 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:10Z","lastTransitionTime":"2025-11-26T06:26:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.100170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.100232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.100250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.100276 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.100294 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.203422 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.203486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.203503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.203529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.203552 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.307175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.307251 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.307270 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.307297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.307319 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.327555 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.327637 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.327643 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.327551 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:11 crc kubenswrapper[4775]: E1126 06:26:11.327796 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:11 crc kubenswrapper[4775]: E1126 06:26:11.327919 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:11 crc kubenswrapper[4775]: E1126 06:26:11.328062 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:11 crc kubenswrapper[4775]: E1126 06:26:11.328194 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.410201 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.410264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.410281 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.410304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.410322 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.513691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.513783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.513801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.513827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.513846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.617505 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.617556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.617572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.617597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.617614 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.720697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.720770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.720784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.720804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.720820 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.825938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.826034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.826063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.826098 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.826135 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.929541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.929597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.929615 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.929639 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:11 crc kubenswrapper[4775]: I1126 06:26:11.929656 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:11Z","lastTransitionTime":"2025-11-26T06:26:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.032085 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.032120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.032132 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.032147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.032158 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.134978 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.135024 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.135035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.135053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.135066 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.237246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.237287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.237297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.237313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.237323 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.340045 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.340107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.340129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.340160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.340180 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.442999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.443056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.443073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.443095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.443116 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.546153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.546204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.546219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.546281 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.546300 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.648925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.648990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.649007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.649035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.649054 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.752328 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.752401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.752426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.752455 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.752477 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.854911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.854979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.855003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.855030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.855052 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.958826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.958900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.958923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.958954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:12 crc kubenswrapper[4775]: I1126 06:26:12.958977 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:12Z","lastTransitionTime":"2025-11-26T06:26:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.062136 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.062273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.062292 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.062316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.062333 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.166110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.166174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.166192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.166219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.166240 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.269133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.269196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.269216 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.269243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.269260 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.327533 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.327658 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.327701 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.327776 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.327954 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.328046 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.328173 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.328293 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.371751 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.371848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.371902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.371927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.371945 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.474917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.474954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.474964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.474980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.474991 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.577565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.577616 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.577633 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.577658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.577677 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.681352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.681416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.681435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.681464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.681484 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.784379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.784448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.784471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.784500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.784524 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.824768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.824814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.824826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.824844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.824859 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.845634 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:13Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.851530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.851579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.851597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.851621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.851638 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.874702 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:13Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.883590 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.883686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.883745 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.883779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.883804 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.904515 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:13Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.909630 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.909680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.909693 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.909738 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.909754 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.927809 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:13Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.933168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.933254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.933279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.933313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.933337 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.957150 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:13Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:13 crc kubenswrapper[4775]: E1126 06:26:13.957377 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.959685 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.959896 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.959927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.959960 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:13 crc kubenswrapper[4775]: I1126 06:26:13.959986 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:13Z","lastTransitionTime":"2025-11-26T06:26:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.063088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.063140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.063157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.063182 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.063201 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.165873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.165954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.165977 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.166002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.166020 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.269067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.269150 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.269178 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.269211 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.269236 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.372665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.372784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.372814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.372847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.372875 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.476091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.476154 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.476172 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.476196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.476213 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.580013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.580074 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.580092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.580116 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.580137 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.683425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.683485 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.683502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.683524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.683542 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.786585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.786656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.786681 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.786709 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.786767 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.889641 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.889688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.889705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.889765 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.889783 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.992829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.992913 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.992936 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.992969 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:14 crc kubenswrapper[4775]: I1126 06:26:14.992990 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:14Z","lastTransitionTime":"2025-11-26T06:26:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.096442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.096492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.096510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.096532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.096549 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.199327 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.199393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.199410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.199436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.199455 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.302552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.302610 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.302627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.302650 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.302668 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.327358 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.327479 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.327384 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.327653 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:15 crc kubenswrapper[4775]: E1126 06:26:15.327629 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:15 crc kubenswrapper[4775]: E1126 06:26:15.327831 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:15 crc kubenswrapper[4775]: E1126 06:26:15.327970 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:15 crc kubenswrapper[4775]: E1126 06:26:15.328049 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.406068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.406121 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.406273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.406305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.406324 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.509389 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.509451 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.509469 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.509493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.509510 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.612596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.612669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.612686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.612754 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.612783 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.715963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.716034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.716075 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.716107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.716124 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.819078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.819135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.819151 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.819175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.819193 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.922595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.922658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.922675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.922701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:15 crc kubenswrapper[4775]: I1126 06:26:15.922756 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:15Z","lastTransitionTime":"2025-11-26T06:26:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.029449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.030263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.030298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.030320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.030342 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.134097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.134162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.134184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.134208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.134225 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.237243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.237298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.237315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.237338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.237355 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.341031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.341124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.341144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.341172 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.341190 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.444424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.444516 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.444552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.444588 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.444612 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.548397 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.548475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.548501 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.548532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.548557 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.651988 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.652051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.652068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.652092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.652108 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.755068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.755143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.755167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.755196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.755220 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.858425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.858489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.858511 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.858540 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.858563 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.961304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.961361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.961379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.961401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:16 crc kubenswrapper[4775]: I1126 06:26:16.961418 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:16Z","lastTransitionTime":"2025-11-26T06:26:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.064557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.064668 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.064692 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.064750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.064777 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.167447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.167575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.167601 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.167627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.167645 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.270494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.270569 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.270587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.271116 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.271184 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.326868 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.326958 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.327004 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:17 crc kubenswrapper[4775]: E1126 06:26:17.327252 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.327518 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:17 crc kubenswrapper[4775]: E1126 06:26:17.327636 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:17 crc kubenswrapper[4775]: E1126 06:26:17.328001 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:17 crc kubenswrapper[4775]: E1126 06:26:17.328183 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.352675 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.373289 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.374575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.374838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.374992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.375126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.375247 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.388141 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.407379 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.424100 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.445085 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.462244 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.478039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.478105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.478129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.478161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.478185 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.521128 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.540222 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.578255 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.581168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.581225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.581243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.581266 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.581284 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.603250 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.620407 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.643459 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.661075 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.677706 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.683496 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.683554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.683572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.683598 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.683616 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.700084 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.721332 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:17Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.785831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.785892 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.785904 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.785921 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.785933 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.888802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.888864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.888885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.888909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.888927 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.991708 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.991805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.991824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.991850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:17 crc kubenswrapper[4775]: I1126 06:26:17.991868 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:17Z","lastTransitionTime":"2025-11-26T06:26:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.095615 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.096098 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.096282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.096445 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.096660 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.199455 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.199490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.199498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.199534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.199543 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.303267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.303332 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.303349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.303375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.303396 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.406211 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.406268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.406286 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.406310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.406329 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.509923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.509979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.509996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.510020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.510037 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.613298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.613798 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.614004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.614179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.614331 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.717534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.717576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.717584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.717597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.717606 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.819868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.820213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.820365 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.820521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.820662 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.923878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.924293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.924455 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.924617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:18 crc kubenswrapper[4775]: I1126 06:26:18.924834 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:18Z","lastTransitionTime":"2025-11-26T06:26:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.028473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.028524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.028543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.028568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.028584 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.131873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.131927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.131943 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.131966 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.131982 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.235990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.236066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.236088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.236113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.236131 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.326899 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.326909 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.327053 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:19 crc kubenswrapper[4775]: E1126 06:26:19.327215 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:19 crc kubenswrapper[4775]: E1126 06:26:19.327379 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:19 crc kubenswrapper[4775]: E1126 06:26:19.327498 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.328471 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:19 crc kubenswrapper[4775]: E1126 06:26:19.328640 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.338933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.338989 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.339006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.339027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.339045 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.441868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.441939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.441950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.441967 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.441979 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.548438 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.548754 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.549665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.549750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.549778 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.653164 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.653229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.653245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.653268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.653286 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.756963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.757030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.757048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.757076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.757095 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.859842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.859903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.859919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.859942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.859961 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.962895 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.962978 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.963001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.963031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:19 crc kubenswrapper[4775]: I1126 06:26:19.963051 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:19Z","lastTransitionTime":"2025-11-26T06:26:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.066479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.066523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.066539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.066562 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.066578 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.170496 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.170567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.170587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.170615 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.170635 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.274524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.274590 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.274613 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.274642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.274663 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.328000 4775 scope.go:117] "RemoveContainer" containerID="023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297" Nov 26 06:26:20 crc kubenswrapper[4775]: E1126 06:26:20.328312 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.377819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.377897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.377919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.377950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.377972 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.480914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.481279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.481450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.481643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.481890 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.585587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.585672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.585691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.585755 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.585776 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.689360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.689426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.689448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.689477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.689495 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.792883 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.792950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.792972 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.793007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.793028 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.896239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.896302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.896325 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.896357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:20 crc kubenswrapper[4775]: I1126 06:26:20.896377 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:20Z","lastTransitionTime":"2025-11-26T06:26:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.000304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.000375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.000392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.000421 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.000441 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.104225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.104294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.104311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.104344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.104367 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.211701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.211781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.211979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.212015 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.212033 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.315788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.315846 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.315864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.315887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.315904 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.327383 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.327444 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.327454 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:21 crc kubenswrapper[4775]: E1126 06:26:21.327508 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.327591 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:21 crc kubenswrapper[4775]: E1126 06:26:21.327681 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:21 crc kubenswrapper[4775]: E1126 06:26:21.327821 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:21 crc kubenswrapper[4775]: E1126 06:26:21.327992 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.419245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.419309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.419329 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.419357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.419376 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.521825 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.521871 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.521884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.521900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.521912 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.624934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.625009 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.625033 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.625062 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.625084 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.727273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.727310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.727321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.727336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.727347 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.838216 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.838254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.838264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.838278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.838286 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.941816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.941881 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.941899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.941924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:21 crc kubenswrapper[4775]: I1126 06:26:21.941941 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:21Z","lastTransitionTime":"2025-11-26T06:26:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.044701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.044806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.044824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.044854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.044872 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.148238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.148286 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.148298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.148315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.148326 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.251435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.251475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.251484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.251501 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.251512 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.354630 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.354872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.354911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.354943 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.354972 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.457071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.457109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.457121 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.457138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.457151 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.559362 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.559387 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.559395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.559409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.559417 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.662998 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.663056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.663074 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.663099 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.663116 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.765068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.765109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.765119 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.765133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.765141 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.867779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.867819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.867829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.867845 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.867854 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.971686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.971793 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.971814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.971839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:22 crc kubenswrapper[4775]: I1126 06:26:22.971856 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:22Z","lastTransitionTime":"2025-11-26T06:26:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.074577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.074660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.074680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.074705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.074760 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.177514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.177573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.177591 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.177618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.177636 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.280437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.280494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.280507 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.280529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.280542 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.327315 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.327414 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.327455 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:23 crc kubenswrapper[4775]: E1126 06:26:23.327629 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.327665 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:23 crc kubenswrapper[4775]: E1126 06:26:23.327781 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:23 crc kubenswrapper[4775]: E1126 06:26:23.327893 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:23 crc kubenswrapper[4775]: E1126 06:26:23.328046 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.383007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.383046 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.383057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.383074 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.383087 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.486015 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.486057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.486066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.486082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.486093 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.589169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.589225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.589237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.589254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.589264 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.692441 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.692545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.692573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.692603 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.692636 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.796013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.796058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.796068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.796082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.796092 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.899343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.899412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.899429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.899456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:23 crc kubenswrapper[4775]: I1126 06:26:23.899475 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:23Z","lastTransitionTime":"2025-11-26T06:26:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.002425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.002465 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.002474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.002490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.002500 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.045030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.045099 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.045123 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.045155 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.045179 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.057369 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:24Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.061988 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.062016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.062025 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.062037 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.062047 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.073485 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:24Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.076910 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.076933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.076942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.076954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.076962 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.088349 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:24Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.092667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.092696 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.092747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.092762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.092772 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.106463 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:24Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.109736 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.109761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.109770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.109783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.109792 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.122221 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:24Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.122320 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.123648 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.123792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.123820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.123852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.123880 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.226039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.226101 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.226120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.226146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.226166 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.230557 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.230680 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:24 crc kubenswrapper[4775]: E1126 06:26:24.230754 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:26:56.230736315 +0000 UTC m=+99.592040267 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.328932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.328993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.329015 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.329042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.329065 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.431096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.431133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.431144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.431161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.431172 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.534066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.534097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.534105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.534118 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.534128 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.636364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.636402 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.636413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.636429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.636440 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.737950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.738021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.738040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.738064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.738098 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.840612 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.840664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.840680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.840704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.840761 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.943242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.943315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.943337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.943368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:24 crc kubenswrapper[4775]: I1126 06:26:24.943391 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:24Z","lastTransitionTime":"2025-11-26T06:26:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.046627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.046697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.046759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.046788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.046813 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.149567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.149623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.149639 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.149661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.149679 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.251511 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.251561 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.251572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.251593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.251607 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.327225 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.327297 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.327385 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:25 crc kubenswrapper[4775]: E1126 06:26:25.327383 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:25 crc kubenswrapper[4775]: E1126 06:26:25.327496 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:25 crc kubenswrapper[4775]: E1126 06:26:25.327577 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.327738 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:25 crc kubenswrapper[4775]: E1126 06:26:25.327803 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.354833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.354869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.354879 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.354892 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.354904 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.457818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.457886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.457924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.457981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.458004 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.560723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.560781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.560791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.560807 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.560818 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.663789 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.663822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.663833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.663848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.663857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.767230 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.767303 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.767320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.767346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.767364 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.826916 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/0.log" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.827032 4775 generic.go:334] "Generic (PLEG): container finished" podID="4181bfb8-478b-4b30-a2cb-744da1010aeb" containerID="43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645" exitCode=1 Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.827085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerDied","Data":"43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.827795 4775 scope.go:117] "RemoveContainer" containerID="43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.841284 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.851082 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.865928 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.869565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.869617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.869633 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.869658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.869674 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.877618 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.902028 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.919463 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.933212 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.948381 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.966439 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.971675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.971710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.971720 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.971739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.971765 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:25Z","lastTransitionTime":"2025-11-26T06:26:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.984175 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:25 crc kubenswrapper[4775]: I1126 06:26:25.995518 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:25Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.010298 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.024533 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.037167 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.048546 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.060127 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.071506 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.073926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.073969 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.073981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.073999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.074011 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.175701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.175755 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.175764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.175780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.175790 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.277297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.277320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.277328 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.277341 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.277349 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.380522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.380889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.380902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.380920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.380931 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.482779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.482809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.482819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.482831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.482840 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.584782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.584856 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.584890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.584915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.584933 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.687467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.687525 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.687544 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.687566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.687586 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.789842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.789901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.789919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.789942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.789957 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.833125 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/0.log" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.833179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerStarted","Data":"a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.849275 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.868943 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.892088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.892130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.892148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.892170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.892186 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.897502 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.919440 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.934322 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.951962 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.966453 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.985048 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:26Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.995413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.995456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.995471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.995490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:26 crc kubenswrapper[4775]: I1126 06:26:26.995505 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:26Z","lastTransitionTime":"2025-11-26T06:26:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.005323 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.016310 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.026173 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.035201 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.042501 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.050239 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.058399 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.069098 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.083518 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.098405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.098440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.098450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.098465 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.098476 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.200905 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.200963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.200982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.201006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.201022 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.303489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.303539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.303552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.303570 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.303583 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.327145 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.327185 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.327276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:27 crc kubenswrapper[4775]: E1126 06:26:27.327624 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.327685 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:27 crc kubenswrapper[4775]: E1126 06:26:27.327763 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:27 crc kubenswrapper[4775]: E1126 06:26:27.327865 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:27 crc kubenswrapper[4775]: E1126 06:26:27.327502 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.341483 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.343224 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.359421 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.374040 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.389864 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.405237 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.405463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.405513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.405530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.405556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.405585 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.419163 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.434479 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.447644 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.463458 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.481927 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.509009 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.509056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.509079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.509156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.509174 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.513314 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.530926 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.542877 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.557314 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.570337 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.583250 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.597495 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:27Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.613816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.613920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.613938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.613961 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.613980 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.717786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.717864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.717889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.717919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.717941 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.820719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.820757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.820784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.820797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.820806 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.923078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.923113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.923124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.923140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:27 crc kubenswrapper[4775]: I1126 06:26:27.923151 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:27Z","lastTransitionTime":"2025-11-26T06:26:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.025521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.025555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.025564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.025582 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.025591 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.128662 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.128707 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.128723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.128771 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.128784 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.264411 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.264452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.264463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.264479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.264491 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.366546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.366593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.366606 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.366622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.366633 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.469097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.469152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.469168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.469195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.469240 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.572786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.572842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.572860 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.572884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.572901 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.675272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.675320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.675336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.675358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.675376 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.777479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.777524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.777541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.777565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.777583 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.880360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.880401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.880409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.880428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.880438 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.982964 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.983007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.983016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.983030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:28 crc kubenswrapper[4775]: I1126 06:26:28.983037 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:28Z","lastTransitionTime":"2025-11-26T06:26:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.086154 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.086218 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.086237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.086264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.086281 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.188639 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.188678 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.188689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.188704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.188718 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.296258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.296558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.296624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.296694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.296842 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.327614 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.327687 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:29 crc kubenswrapper[4775]: E1126 06:26:29.327743 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.327628 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.327791 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:29 crc kubenswrapper[4775]: E1126 06:26:29.327899 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:29 crc kubenswrapper[4775]: E1126 06:26:29.328009 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:29 crc kubenswrapper[4775]: E1126 06:26:29.328046 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.399205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.399246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.399256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.399271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.399282 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.501549 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.501787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.501890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.501988 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.502072 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.604710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.604826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.604847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.604877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.604897 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.708258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.708504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.708658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.708754 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.708857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.812494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.812548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.812565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.812589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.812606 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.916159 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.916202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.916216 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.916235 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:29 crc kubenswrapper[4775]: I1126 06:26:29.916250 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:29Z","lastTransitionTime":"2025-11-26T06:26:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.019016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.019090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.019105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.019124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.019162 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.121641 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.121699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.121752 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.121776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.121794 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.226979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.227087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.227106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.227130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.227146 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.329864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.329931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.329948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.329972 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.329989 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.432405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.432462 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.432479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.432503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.432522 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.535238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.535305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.535328 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.535354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.535373 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.638564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.638617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.638633 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.638655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.638672 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.742060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.742148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.742181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.742215 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.742238 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.844652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.844761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.844787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.844815 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.844839 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.947973 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.948043 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.948065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.948087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:30 crc kubenswrapper[4775]: I1126 06:26:30.948104 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:30Z","lastTransitionTime":"2025-11-26T06:26:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.050207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.050304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.050321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.050343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.050360 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.153818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.153891 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.153916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.153945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.153967 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.257111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.257192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.257214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.257246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.257269 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.326766 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.326833 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.326851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.326784 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:31 crc kubenswrapper[4775]: E1126 06:26:31.326990 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:31 crc kubenswrapper[4775]: E1126 06:26:31.327105 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:31 crc kubenswrapper[4775]: E1126 06:26:31.327295 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:31 crc kubenswrapper[4775]: E1126 06:26:31.327410 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.360108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.360174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.360195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.360223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.360246 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.462963 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.463023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.463041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.463064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.463082 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.575920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.576007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.576026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.576048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.576066 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.678946 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.679005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.679022 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.679045 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.679063 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.782331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.782385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.782401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.782425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.782446 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.885940 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.885996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.886012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.886034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.886055 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.989114 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.989160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.989180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.989203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:31 crc kubenswrapper[4775]: I1126 06:26:31.989222 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:31Z","lastTransitionTime":"2025-11-26T06:26:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.092305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.092401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.092426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.092456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.092479 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.195090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.195140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.195157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.195181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.195197 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.298757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.298844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.298862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.298882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.298935 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.401151 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.401208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.401227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.401250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.401266 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.504330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.504393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.504413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.504439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.504457 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.607566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.607626 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.607643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.607668 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.607685 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.710853 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.710968 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.710996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.711022 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.711044 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.813504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.813568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.813585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.813610 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.813631 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.916541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.916621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.916643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.916672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:32 crc kubenswrapper[4775]: I1126 06:26:32.916694 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:32Z","lastTransitionTime":"2025-11-26T06:26:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.019992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.020041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.020052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.020072 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.020085 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.125358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.125417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.125427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.125444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.125455 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.228414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.228490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.228516 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.228547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.228570 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.327327 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.327427 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.327423 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.327502 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:33 crc kubenswrapper[4775]: E1126 06:26:33.327475 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:33 crc kubenswrapper[4775]: E1126 06:26:33.327771 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:33 crc kubenswrapper[4775]: E1126 06:26:33.327829 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:33 crc kubenswrapper[4775]: E1126 06:26:33.327907 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.329106 4775 scope.go:117] "RemoveContainer" containerID="023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.331345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.331376 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.331387 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.331404 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.331416 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.434005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.434079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.434102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.434149 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.434174 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.537111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.537164 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.537188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.537216 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.537238 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.640367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.640448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.640467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.640494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.640510 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.742928 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.742965 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.742976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.742992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.743004 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.846449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.846496 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.846513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.846539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.846556 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.856170 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/2.log" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.859459 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.859983 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.877144 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.903960 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebb1d370-8dfe-46c2-bb5a-f77fe8c74fd6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d9cb9007f907be0090f6b917c6a58f654971b43b1ff4d278d5787f1852c2ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.916600 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.932486 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.948874 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.948925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.948938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.948956 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.948969 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:33Z","lastTransitionTime":"2025-11-26T06:26:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.951122 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.972757 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:33 crc kubenswrapper[4775]: I1126 06:26:33.997907 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:33Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.013217 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.028956 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.044000 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.051866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.051919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.051936 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.051960 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.051977 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.059823 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.069194 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.078534 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.089849 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.102107 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.119028 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.134199 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.149242 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.154119 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.154179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.154198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.154228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.154248 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.257219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.257287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.257305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.257330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.257348 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.360232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.360280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.360298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.360321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.360339 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.463460 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.463527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.463544 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.463567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.463587 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.477400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.477466 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.477486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.477513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.477534 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.500258 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.505261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.505354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.505408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.505432 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.505451 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.528149 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.533298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.533358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.533376 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.533400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.533419 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.554055 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.558509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.558538 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.558548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.558565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.558579 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.578534 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.583348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.583448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.583501 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.583528 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.583547 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.603895 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ca7ec27e-1a29-480c-b7d0-2bfb73b424b6\\\",\\\"systemUUID\\\":\\\"7c9a1c38-84d7-4412-b12e-7045594af10c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.604111 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.606795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.606857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.606882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.606911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.606930 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.709693 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.709760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.709797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.709820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.709838 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.812622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.812698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.812762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.812789 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.812806 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.866205 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/3.log" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.867545 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/2.log" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.871972 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" exitCode=1 Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.872022 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.872076 4775 scope.go:117] "RemoveContainer" containerID="023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.873282 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:26:34 crc kubenswrapper[4775]: E1126 06:26:34.873599 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.893403 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.916404 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.917339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.917385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.917402 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.917425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.917442 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:34Z","lastTransitionTime":"2025-11-26T06:26:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.938756 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.958400 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:34 crc kubenswrapper[4775]: I1126 06:26:34.978833 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.000248 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:34Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.017174 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.020530 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.020572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.020589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.020660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.020678 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.031961 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.050048 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.066532 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.087957 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.108868 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.123198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.123254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.123271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.123294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.123311 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.129022 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.149246 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.180108 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://023efb47b4fb25325da6264908ba719d3afdf5a75e64704dec3a0b8bd5be2297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:08Z\\\",\\\"message\\\":\\\" 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.Namespace\\\\nI1126 06:26:08.309067 6415 egressqos.go:301] Shutting down EgressQoS controller\\\\nI1126 06:26:08.309117 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressFirewall\\\\nI1126 06:26:08.309143 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.EgressIP\\\\nI1126 06:26:08.309162 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressNode\\\\nI1126 06:26:08.309197 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPPod\\\\nI1126 06:26:08.309217 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *factory.egressIPNamespace\\\\nI1126 06:26:08.309239 6415 obj_retry.go:439] Stop channel got triggered: will stop retrying failed objects of type *v1.NetworkPolicy\\\\nI1126 06:26:08.309333 6415 nad_controller.go:166] [zone-nad-controller NAD controller]: shutting down\\\\nI1126 06:26:08.315811 6415 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1126 06:26:08.315939 6415 factory.go:656] Stopping watch factory\\\\nI1126 06:26:08.315961 6415 ovnkube.go:599] Stopped ovnkube\\\\nI1126 06:26:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"ult, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.139\\\\\\\", Port:17698, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1126 06:26:34.246933 6769 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.203317 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.218980 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebb1d370-8dfe-46c2-bb5a-f77fe8c74fd6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d9cb9007f907be0090f6b917c6a58f654971b43b1ff4d278d5787f1852c2ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.225759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.225826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.225850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.225878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.225899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.235704 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.327145 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.327278 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.327211 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.327651 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:35 crc kubenswrapper[4775]: E1126 06:26:35.327988 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:35 crc kubenswrapper[4775]: E1126 06:26:35.328138 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:35 crc kubenswrapper[4775]: E1126 06:26:35.328270 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:35 crc kubenswrapper[4775]: E1126 06:26:35.328448 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.328968 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.329021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.329038 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.329061 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.329084 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.431332 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.431365 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.431373 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.431385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.431394 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.534802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.534871 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.534891 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.534918 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.534938 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.638274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.638367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.638383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.638409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.638426 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.741417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.741476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.741486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.741500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.741509 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.844026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.844130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.844148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.844176 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.844195 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.878804 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/3.log" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.884844 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:26:35 crc kubenswrapper[4775]: E1126 06:26:35.885121 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.901404 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.918992 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.934929 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.948468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.948524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.948542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.948565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.948582 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:35Z","lastTransitionTime":"2025-11-26T06:26:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.958332 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.975560 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:35 crc kubenswrapper[4775]: I1126 06:26:35.995227 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:35Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.012226 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.043678 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"ult, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.139\\\\\\\", Port:17698, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1126 06:26:34.246933 6769 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.051864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.051914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.051932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.051955 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.051972 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.071571 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.090504 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebb1d370-8dfe-46c2-bb5a-f77fe8c74fd6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d9cb9007f907be0090f6b917c6a58f654971b43b1ff4d278d5787f1852c2ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.105700 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.126613 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.147103 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.156870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.156960 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.156978 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.157045 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.157064 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.168838 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.189275 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.209511 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.227064 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.244887 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:36Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.260249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.260316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.260339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.260368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.260389 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.363769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.363837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.363858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.363882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.363899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.466916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.466982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.466999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.467023 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.467040 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.570192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.570288 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.570307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.570345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.570382 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.673177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.673228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.673252 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.673285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.673305 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.776047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.776112 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.776133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.776161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.776180 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.885214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.885301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.885322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.885354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.885376 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.989264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.989324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.989344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.989372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:36 crc kubenswrapper[4775]: I1126 06:26:36.989392 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:36Z","lastTransitionTime":"2025-11-26T06:26:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.092599 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.092656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.092675 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.092706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.092759 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.196321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.196935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.196954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.196983 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.197002 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.299270 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.299319 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.299337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.299362 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.299380 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.327104 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.327123 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:37 crc kubenswrapper[4775]: E1126 06:26:37.327229 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.327116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.327309 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:37 crc kubenswrapper[4775]: E1126 06:26:37.327533 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:37 crc kubenswrapper[4775]: E1126 06:26:37.327677 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:37 crc kubenswrapper[4775]: E1126 06:26:37.327966 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.344481 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5739dc36-9102-437f-95c0-b9e137f4a458\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fb08f20b936a5a45757586d14087f2a74c384a9845d1160c32c29661548bdc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca3dc8b8cedef298a72d1ce19825207faa58d9dabe7499e83d9157334c8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2hcx9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:50Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-tmqbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.357089 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n98r8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e24b0f5-8555-40a4-a965-e67d5cd305a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jz7rv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n98r8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.378640 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2940e7dc-6217-45e9-9aaf-b889adf5058a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T06:25:36Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 06:25:30.848020 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 06:25:30.852181 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2107676991/tls.crt::/tmp/serving-cert-2107676991/tls.key\\\\\\\"\\\\nI1126 06:25:36.636878 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 06:25:36.641442 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 06:25:36.641542 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 06:25:36.641622 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 06:25:36.641668 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 06:25:36.649193 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 06:25:36.649240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649252 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 06:25:36.649261 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 06:25:36.649268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 06:25:36.649274 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 06:25:36.649280 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 06:25:36.649209 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 06:25:36.662255 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.397158 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.401887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.401950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.401974 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.402007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.402031 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.413082 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.427059 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"687d9603-b248-4ad7-95d6-a23c110afaf1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92ed74c8a23c4edac9526773d61a3fa7496e5c856326558dc97e5d855d24ba7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dx67z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-p7z2d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.437252 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-456fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa50ee33-2e13-4179-bfc4-37cad702ef1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d78560b2d56265b5e1fac6bd6d9a490d050fe2915307fd2db8e4b5ceb48c736e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zdtfx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-456fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.455919 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e091b0f3-e906-47c0-93a5-5df598bc954e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba439631f5c6e6949295a5cc1c75c5b01ef5b6982cedfdfcd028748c760a4f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b58fec91b078aa9bdd88fc5cc7afa35dc53499eca9a0be620c57197dc255c6a5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://147b67d3be6e315625259372481090bcdefd11c4664e78a4ef0c668aaef7901a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ea4df8d8d8051c54ce44a4e6ac550b00783877a627cc5f0183cee39c3b627446\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b662b40b886b8dca56726a0bfe25ba8b1b1b2a0f0e0fff719a5abc5cdc67c4e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bfcea976943f426da57403e7ed8583f1bb6bf1ced7815af51608b97ad2f0e173\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85edf99d4384cb643e0810dbfef92daf4922bb62bd610fdba5012f8f523ca135\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bjzj5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zbl4m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.468847 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebb1d370-8dfe-46c2-bb5a-f77fe8c74fd6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d9cb9007f907be0090f6b917c6a58f654971b43b1ff4d278d5787f1852c2ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b16623ba88cfa69ddc288782b58f06ad7d295c4b3cd3d19ae063ca2fb47f9b10\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.479486 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t5r7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5458a69d-5519-4863-baf9-e549e5e52b78\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0eee58f562621f5f52baa105541430556ab73d704da55a020fc32c63bbe9fcb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bn4jr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t5r7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.492675 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfbe13fc5cae2831dc2d601c0a9658fd10cb876f24e941141b94d8d47c3e775\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.504550 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.504642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.504759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.504800 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.504821 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.505428 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47b87c33e5ec6be665993371f50dff1583fb763bb689a36e6aa6f8d290deb29d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43e23cd2fea4e0088eb9b2eaa43ed0a64104b8615277f15a3ae5d91b4aca214c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.518579 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k9pcc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4181bfb8-478b-4b30-a2cb-744da1010aeb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:25Z\\\",\\\"message\\\":\\\"2025-11-26T06:25:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989\\\\n2025-11-26T06:25:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_37dca6dc-76d0-4050-8c34-17c58a9fe989 to /host/opt/cni/bin/\\\\n2025-11-26T06:25:40Z [verbose] multus-daemon started\\\\n2025-11-26T06:25:40Z [verbose] Readiness Indicator file check\\\\n2025-11-26T06:26:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:26:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srhx4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k9pcc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.541877 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"53d5d732-edab-4a9d-802c-4d9f0a2197ac\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T06:26:34Z\\\",\\\"message\\\":\\\"ult, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"8efa4d1a-72f5-4dfa-9bc2-9d93ef11ecf2\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-apiserver/check-endpoints_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/check-endpoints\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.139\\\\\\\", Port:17698, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1126 06:26:34.246933 6769 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T06:26:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4g4l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f9lb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.557168 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3f9e7812-ea50-4cc5-a807-8d12a481b658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dff3790d2692231d2776a963a252a38d5d35c6954e081f49fb00981b5f81b678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f32f3c28dbe2f174ba8e20a458038a5c7d431ec727f1ebc5b1c97983aa85f1fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5f1fbfef3ceb7238cfda9d05c2c3e89f1a219d234ea95ddd9b9047e14d22d0a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.574642 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.592475 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b61fcfda46263ca8464b8d0e9dc6b80dbb456a10efe9b4e3efe91905087ac203\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.608159 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.608201 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.608213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.608229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.608242 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.608629 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb7c9424-0d36-483d-8e1e-52389cea0115\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:26:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T06:25:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f025b543fc9c32626cb60abf87991a5ae33a0650826d4fa5d74f5b3cd53d8f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79e4d443b55ba67909a7092c7126097dc7d4070b6acd2287a30efd97b483e379\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72b281a96f113f90600f0a823a0f40cd680dbb60f5d6915d9334f14bc108f614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T06:25:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8468c2f295eeab4d5e3cf3bb8c1b6d829675a510f34d9e0de88d2c76f5fe6d4b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T06:25:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T06:25:18Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T06:25:17Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T06:26:37Z is after 2025-08-24T17:21:41Z" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.711383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.711442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.711458 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.711480 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.711497 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.814228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.814283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.814299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.814321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.814340 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.917035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.917092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.917108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.917130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:37 crc kubenswrapper[4775]: I1126 06:26:37.917146 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:37Z","lastTransitionTime":"2025-11-26T06:26:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.019950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.020031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.020067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.020100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.020117 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.122434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.122494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.122513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.122535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.122553 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.225954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.225990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.226006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.226026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.226043 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.349638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.350071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.350274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.350595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.351313 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.455368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.455424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.455631 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.455659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.455672 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.558187 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.558238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.558249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.558267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.558279 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.661261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.661315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.661332 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.661355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.661372 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.763977 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.764033 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.764050 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.764066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.764077 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.866435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.866493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.866503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.866522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.866534 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.968598 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.968640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.968652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.968667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:38 crc kubenswrapper[4775]: I1126 06:26:38.968676 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:38Z","lastTransitionTime":"2025-11-26T06:26:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.072263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.072331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.072353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.072383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.072405 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.175096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.175145 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.175161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.175184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.175201 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.278916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.278994 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.279017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.279047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.279065 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.327935 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.327915 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.327977 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:39 crc kubenswrapper[4775]: E1126 06:26:39.328209 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:39 crc kubenswrapper[4775]: E1126 06:26:39.328360 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.328421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:39 crc kubenswrapper[4775]: E1126 06:26:39.328526 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:39 crc kubenswrapper[4775]: E1126 06:26:39.328655 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.382207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.382265 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.382282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.382306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.382323 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.485228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.485289 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.485306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.485330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.485348 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.588740 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.588814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.588875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.588944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.588982 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.692158 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.692238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.692262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.692343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.692368 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.796310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.796378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.796395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.796419 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.796438 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.900497 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.900565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.900582 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.900607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:39 crc kubenswrapper[4775]: I1126 06:26:39.900623 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:39Z","lastTransitionTime":"2025-11-26T06:26:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.004427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.004487 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.004503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.004526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.004546 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.108076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.108124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.108140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.108162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.108178 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.211012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.211088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.211109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.211137 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.211154 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.313868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.313936 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.313954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.313980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.313999 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.417233 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.417300 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.417316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.417339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.417359 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.520831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.520929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.520954 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.520984 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.521004 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.624856 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.624937 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.624958 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.625428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.625488 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.729035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.729101 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.729118 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.729142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.729158 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.832144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.832205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.832222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.832243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.832260 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.934493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.934552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.934581 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.934610 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:40 crc kubenswrapper[4775]: I1126 06:26:40.934637 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:40Z","lastTransitionTime":"2025-11-26T06:26:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.037761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.037833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.037867 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.037896 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.037916 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.121407 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.121608 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.121680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.121891 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.121972 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:27:45.121945189 +0000 UTC m=+148.483249181 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.122054 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:45.122015961 +0000 UTC m=+148.483319953 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.122141 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.122226 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 06:27:45.122205215 +0000 UTC m=+148.483509247 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.140950 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.141000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.141017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.141044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.141063 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.223272 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.223427 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223559 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223612 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223617 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223643 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223656 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223669 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223803 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 06:27:45.223780169 +0000 UTC m=+148.585084161 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.223833 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 06:27:45.223821141 +0000 UTC m=+148.585125133 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.244107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.244160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.244177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.244200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.244217 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.327273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.327338 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.327464 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.327466 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.327539 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.327709 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.327880 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:41 crc kubenswrapper[4775]: E1126 06:26:41.327985 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.346705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.346786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.346805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.346839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.346857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.450120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.450184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.450202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.450226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.450243 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.553410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.553494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.553526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.553553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.553573 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.657027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.657133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.657146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.657161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.657172 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.761089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.761129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.761139 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.761178 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.761189 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.864065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.864358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.864577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.864762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.864917 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.968560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.968613 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.968629 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.968651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:41 crc kubenswrapper[4775]: I1126 06:26:41.968669 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:41Z","lastTransitionTime":"2025-11-26T06:26:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.071513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.071563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.071578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.071601 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.071618 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.175204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.175558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.175776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.175933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.176077 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.279656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.279762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.279783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.279810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.279831 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.383574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.383645 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.383663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.383688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.383710 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.489202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.489236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.489248 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.489270 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.489284 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.591797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.591883 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.591903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.591968 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.591985 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.694852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.694926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.694951 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.694983 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.695004 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.797584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.797648 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.797666 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.797691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.797745 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.900909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.900980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.900997 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.901022 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:42 crc kubenswrapper[4775]: I1126 06:26:42.901040 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:42Z","lastTransitionTime":"2025-11-26T06:26:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.003636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.003693 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.003741 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.003773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.003796 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.106066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.106134 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.106160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.106194 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.106219 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.208880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.208979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.208996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.209020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.209036 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.312347 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.312610 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.312809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.312952 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.313109 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.326862 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.326971 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:43 crc kubenswrapper[4775]: E1126 06:26:43.327013 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:43 crc kubenswrapper[4775]: E1126 06:26:43.327147 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.327236 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:43 crc kubenswrapper[4775]: E1126 06:26:43.327329 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.327372 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:43 crc kubenswrapper[4775]: E1126 06:26:43.327487 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.415882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.415932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.415943 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.415966 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.415978 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.518470 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.518512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.518524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.518542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.518553 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.621819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.621894 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.621922 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.621953 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.621977 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.725448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.725519 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.725542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.725569 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.725587 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.829153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.829219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.829236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.829260 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.829286 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.932456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.932520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.932538 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.932561 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:43 crc kubenswrapper[4775]: I1126 06:26:43.932581 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:43Z","lastTransitionTime":"2025-11-26T06:26:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.035788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.035860 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.035886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.035915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.035937 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.139130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.139209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.139228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.139253 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.139272 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.242293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.242352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.242368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.242391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.242409 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.346036 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.346094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.346113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.346138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.346159 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.449644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.449788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.449810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.449833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.449852 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.553357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.553418 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.553456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.553489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.553513 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.657526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.657579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.657600 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.657628 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.657650 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.760844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.760901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.760921 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.760944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.760961 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.789269 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.789328 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.789344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.789367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.789384 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T06:26:44Z","lastTransitionTime":"2025-11-26T06:26:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.867244 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z"] Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.867814 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.873443 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.873837 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.873970 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.874020 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.896538 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=67.896508085 podStartE2EDuration="1m7.896508085s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:44.896083624 +0000 UTC m=+88.257387596" watchObservedRunningTime="2025-11-26 06:26:44.896508085 +0000 UTC m=+88.257812077" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.961495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.961558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.961835 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.961986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.962047 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.964081 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=35.964057673 podStartE2EDuration="35.964057673s" podCreationTimestamp="2025-11-26 06:26:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:44.94606768 +0000 UTC m=+88.307371712" watchObservedRunningTime="2025-11-26 06:26:44.964057673 +0000 UTC m=+88.325361695" Nov 26 06:26:44 crc kubenswrapper[4775]: I1126 06:26:44.982800 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-tmqbx" podStartSLOduration=66.982767685 podStartE2EDuration="1m6.982767685s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:44.982398495 +0000 UTC m=+88.343702487" watchObservedRunningTime="2025-11-26 06:26:44.982767685 +0000 UTC m=+88.344071667" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.028495 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=69.02844569 podStartE2EDuration="1m9.02844569s" podCreationTimestamp="2025-11-26 06:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.027998399 +0000 UTC m=+88.389302371" watchObservedRunningTime="2025-11-26 06:26:45.02844569 +0000 UTC m=+88.389749702" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063435 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063707 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.063786 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.065595 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.073166 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.093352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87e3fdb1-ede1-47ff-829c-6d9c33ba40bf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-kzc9z\" (UID: \"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.095843 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-456fr" podStartSLOduration=68.095828905 podStartE2EDuration="1m8.095828905s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.095792354 +0000 UTC m=+88.457096326" watchObservedRunningTime="2025-11-26 06:26:45.095828905 +0000 UTC m=+88.457132857" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.096013 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podStartSLOduration=68.096008479 podStartE2EDuration="1m8.096008479s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.083103137 +0000 UTC m=+88.444407089" watchObservedRunningTime="2025-11-26 06:26:45.096008479 +0000 UTC m=+88.457312431" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.127228 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zbl4m" podStartSLOduration=68.127208612 podStartE2EDuration="1m8.127208612s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.114730131 +0000 UTC m=+88.476034103" watchObservedRunningTime="2025-11-26 06:26:45.127208612 +0000 UTC m=+88.488512714" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.127410 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=18.127406567 podStartE2EDuration="18.127406567s" podCreationTimestamp="2025-11-26 06:26:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.127024228 +0000 UTC m=+88.488328190" watchObservedRunningTime="2025-11-26 06:26:45.127406567 +0000 UTC m=+88.488710519" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.152329 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-t5r7v" podStartSLOduration=69.152300428 podStartE2EDuration="1m9.152300428s" podCreationTimestamp="2025-11-26 06:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.138255857 +0000 UTC m=+88.499559809" watchObservedRunningTime="2025-11-26 06:26:45.152300428 +0000 UTC m=+88.513604420" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.179940 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-k9pcc" podStartSLOduration=68.179913069 podStartE2EDuration="1m8.179913069s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.178910273 +0000 UTC m=+88.540214225" watchObservedRunningTime="2025-11-26 06:26:45.179913069 +0000 UTC m=+88.541217081" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.193094 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.327202 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.327213 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:45 crc kubenswrapper[4775]: E1126 06:26:45.327389 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.327216 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.327440 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:45 crc kubenswrapper[4775]: E1126 06:26:45.327528 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:45 crc kubenswrapper[4775]: E1126 06:26:45.327603 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:45 crc kubenswrapper[4775]: E1126 06:26:45.327647 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.928265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" event={"ID":"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf","Type":"ContainerStarted","Data":"81f63c98f3cbe9083dbfbf926f7b24a67dd2f1ff58eb08f795dc8892b2fd015c"} Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.928359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" event={"ID":"87e3fdb1-ede1-47ff-829c-6d9c33ba40bf","Type":"ContainerStarted","Data":"1a906ab3403a293103a33781ee4491c62a4ad7c29a2d433a16ffb0af39cec355"} Nov 26 06:26:45 crc kubenswrapper[4775]: I1126 06:26:45.950278 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-kzc9z" podStartSLOduration=68.950251745 podStartE2EDuration="1m8.950251745s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:45.949350752 +0000 UTC m=+89.310654764" watchObservedRunningTime="2025-11-26 06:26:45.950251745 +0000 UTC m=+89.311555737" Nov 26 06:26:47 crc kubenswrapper[4775]: I1126 06:26:47.327031 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:47 crc kubenswrapper[4775]: I1126 06:26:47.327064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:47 crc kubenswrapper[4775]: I1126 06:26:47.327128 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:47 crc kubenswrapper[4775]: I1126 06:26:47.327182 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:47 crc kubenswrapper[4775]: E1126 06:26:47.329222 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:47 crc kubenswrapper[4775]: E1126 06:26:47.329526 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:47 crc kubenswrapper[4775]: E1126 06:26:47.329699 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:47 crc kubenswrapper[4775]: E1126 06:26:47.330053 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:48 crc kubenswrapper[4775]: I1126 06:26:48.356349 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 26 06:26:49 crc kubenswrapper[4775]: I1126 06:26:49.326962 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:49 crc kubenswrapper[4775]: I1126 06:26:49.327047 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:49 crc kubenswrapper[4775]: I1126 06:26:49.327080 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:49 crc kubenswrapper[4775]: I1126 06:26:49.327164 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:49 crc kubenswrapper[4775]: E1126 06:26:49.327160 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:49 crc kubenswrapper[4775]: E1126 06:26:49.327598 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:49 crc kubenswrapper[4775]: E1126 06:26:49.327749 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:49 crc kubenswrapper[4775]: E1126 06:26:49.327860 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:51 crc kubenswrapper[4775]: I1126 06:26:51.326945 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:51 crc kubenswrapper[4775]: I1126 06:26:51.327050 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:51 crc kubenswrapper[4775]: E1126 06:26:51.327174 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:51 crc kubenswrapper[4775]: I1126 06:26:51.327307 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:51 crc kubenswrapper[4775]: I1126 06:26:51.327309 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:51 crc kubenswrapper[4775]: E1126 06:26:51.327420 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:51 crc kubenswrapper[4775]: E1126 06:26:51.327575 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:51 crc kubenswrapper[4775]: E1126 06:26:51.328315 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:51 crc kubenswrapper[4775]: I1126 06:26:51.328961 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:26:51 crc kubenswrapper[4775]: E1126 06:26:51.329268 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:26:53 crc kubenswrapper[4775]: I1126 06:26:53.327972 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:53 crc kubenswrapper[4775]: E1126 06:26:53.328313 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:53 crc kubenswrapper[4775]: I1126 06:26:53.327994 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:53 crc kubenswrapper[4775]: E1126 06:26:53.328383 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:53 crc kubenswrapper[4775]: I1126 06:26:53.328102 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:53 crc kubenswrapper[4775]: E1126 06:26:53.328426 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:53 crc kubenswrapper[4775]: I1126 06:26:53.327972 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:53 crc kubenswrapper[4775]: E1126 06:26:53.328469 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:55 crc kubenswrapper[4775]: I1126 06:26:55.327095 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:55 crc kubenswrapper[4775]: I1126 06:26:55.327144 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:55 crc kubenswrapper[4775]: I1126 06:26:55.327192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:55 crc kubenswrapper[4775]: E1126 06:26:55.327335 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:55 crc kubenswrapper[4775]: I1126 06:26:55.327390 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:55 crc kubenswrapper[4775]: E1126 06:26:55.327529 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:55 crc kubenswrapper[4775]: E1126 06:26:55.327664 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:55 crc kubenswrapper[4775]: E1126 06:26:55.327860 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:56 crc kubenswrapper[4775]: I1126 06:26:56.321944 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:56 crc kubenswrapper[4775]: E1126 06:26:56.322197 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:56 crc kubenswrapper[4775]: E1126 06:26:56.322282 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs podName:1e24b0f5-8555-40a4-a965-e67d5cd305a6 nodeName:}" failed. No retries permitted until 2025-11-26 06:28:00.322257461 +0000 UTC m=+163.683561443 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs") pod "network-metrics-daemon-n98r8" (UID: "1e24b0f5-8555-40a4-a965-e67d5cd305a6") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 06:26:57 crc kubenswrapper[4775]: I1126 06:26:57.327838 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:57 crc kubenswrapper[4775]: I1126 06:26:57.327874 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:57 crc kubenswrapper[4775]: I1126 06:26:57.327928 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:57 crc kubenswrapper[4775]: E1126 06:26:57.329567 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:57 crc kubenswrapper[4775]: I1126 06:26:57.329692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:57 crc kubenswrapper[4775]: E1126 06:26:57.329873 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:57 crc kubenswrapper[4775]: E1126 06:26:57.330080 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:57 crc kubenswrapper[4775]: E1126 06:26:57.330241 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:26:57 crc kubenswrapper[4775]: I1126 06:26:57.371940 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.371913356 podStartE2EDuration="9.371913356s" podCreationTimestamp="2025-11-26 06:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:26:57.37012835 +0000 UTC m=+100.731432352" watchObservedRunningTime="2025-11-26 06:26:57.371913356 +0000 UTC m=+100.733217368" Nov 26 06:26:59 crc kubenswrapper[4775]: I1126 06:26:59.327333 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:26:59 crc kubenswrapper[4775]: I1126 06:26:59.327455 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:26:59 crc kubenswrapper[4775]: I1126 06:26:59.327467 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:26:59 crc kubenswrapper[4775]: I1126 06:26:59.327565 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:26:59 crc kubenswrapper[4775]: E1126 06:26:59.327619 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:26:59 crc kubenswrapper[4775]: E1126 06:26:59.327886 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:26:59 crc kubenswrapper[4775]: E1126 06:26:59.328067 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:26:59 crc kubenswrapper[4775]: E1126 06:26:59.328183 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:01 crc kubenswrapper[4775]: I1126 06:27:01.326859 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:01 crc kubenswrapper[4775]: I1126 06:27:01.326968 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:01 crc kubenswrapper[4775]: E1126 06:27:01.327047 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:01 crc kubenswrapper[4775]: I1126 06:27:01.327076 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:01 crc kubenswrapper[4775]: I1126 06:27:01.326968 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:01 crc kubenswrapper[4775]: E1126 06:27:01.327226 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:01 crc kubenswrapper[4775]: E1126 06:27:01.327389 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:01 crc kubenswrapper[4775]: E1126 06:27:01.327533 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:03 crc kubenswrapper[4775]: I1126 06:27:03.327189 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:03 crc kubenswrapper[4775]: I1126 06:27:03.327256 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:03 crc kubenswrapper[4775]: I1126 06:27:03.327320 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:03 crc kubenswrapper[4775]: E1126 06:27:03.327614 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:03 crc kubenswrapper[4775]: I1126 06:27:03.327650 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:03 crc kubenswrapper[4775]: E1126 06:27:03.327867 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:03 crc kubenswrapper[4775]: E1126 06:27:03.328095 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:03 crc kubenswrapper[4775]: E1126 06:27:03.328192 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:05 crc kubenswrapper[4775]: I1126 06:27:05.326989 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:05 crc kubenswrapper[4775]: E1126 06:27:05.327191 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:05 crc kubenswrapper[4775]: I1126 06:27:05.327278 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:05 crc kubenswrapper[4775]: I1126 06:27:05.327331 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:05 crc kubenswrapper[4775]: I1126 06:27:05.327578 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:05 crc kubenswrapper[4775]: E1126 06:27:05.327767 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:05 crc kubenswrapper[4775]: E1126 06:27:05.327993 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:05 crc kubenswrapper[4775]: E1126 06:27:05.328137 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:06 crc kubenswrapper[4775]: I1126 06:27:06.328278 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:27:06 crc kubenswrapper[4775]: E1126 06:27:06.328528 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f9lb8_openshift-ovn-kubernetes(53d5d732-edab-4a9d-802c-4d9f0a2197ac)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" Nov 26 06:27:07 crc kubenswrapper[4775]: I1126 06:27:07.327624 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:07 crc kubenswrapper[4775]: I1126 06:27:07.327976 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:07 crc kubenswrapper[4775]: E1126 06:27:07.329849 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:07 crc kubenswrapper[4775]: I1126 06:27:07.329912 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:07 crc kubenswrapper[4775]: I1126 06:27:07.329885 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:07 crc kubenswrapper[4775]: E1126 06:27:07.330045 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:07 crc kubenswrapper[4775]: E1126 06:27:07.330163 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:07 crc kubenswrapper[4775]: E1126 06:27:07.330337 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:09 crc kubenswrapper[4775]: I1126 06:27:09.326639 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:09 crc kubenswrapper[4775]: E1126 06:27:09.326897 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:09 crc kubenswrapper[4775]: I1126 06:27:09.326963 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:09 crc kubenswrapper[4775]: I1126 06:27:09.327004 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:09 crc kubenswrapper[4775]: I1126 06:27:09.327021 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:09 crc kubenswrapper[4775]: E1126 06:27:09.327426 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:09 crc kubenswrapper[4775]: E1126 06:27:09.327851 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:09 crc kubenswrapper[4775]: E1126 06:27:09.327997 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:11 crc kubenswrapper[4775]: I1126 06:27:11.326935 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:11 crc kubenswrapper[4775]: I1126 06:27:11.327000 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:11 crc kubenswrapper[4775]: I1126 06:27:11.327064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:11 crc kubenswrapper[4775]: E1126 06:27:11.327246 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:11 crc kubenswrapper[4775]: I1126 06:27:11.327287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:11 crc kubenswrapper[4775]: E1126 06:27:11.327444 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:11 crc kubenswrapper[4775]: E1126 06:27:11.327585 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:11 crc kubenswrapper[4775]: E1126 06:27:11.327667 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:12 crc kubenswrapper[4775]: I1126 06:27:12.025266 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/1.log" Nov 26 06:27:12 crc kubenswrapper[4775]: I1126 06:27:12.026247 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/0.log" Nov 26 06:27:12 crc kubenswrapper[4775]: I1126 06:27:12.026343 4775 generic.go:334] "Generic (PLEG): container finished" podID="4181bfb8-478b-4b30-a2cb-744da1010aeb" containerID="a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94" exitCode=1 Nov 26 06:27:12 crc kubenswrapper[4775]: I1126 06:27:12.026393 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerDied","Data":"a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94"} Nov 26 06:27:12 crc kubenswrapper[4775]: I1126 06:27:12.026447 4775 scope.go:117] "RemoveContainer" containerID="43574c06772bcdd394b39f600979a2c03d73017a3fb6b43b2f78f6ec21051645" Nov 26 06:27:12 crc kubenswrapper[4775]: I1126 06:27:12.027264 4775 scope.go:117] "RemoveContainer" containerID="a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94" Nov 26 06:27:12 crc kubenswrapper[4775]: E1126 06:27:12.027570 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-k9pcc_openshift-multus(4181bfb8-478b-4b30-a2cb-744da1010aeb)\"" pod="openshift-multus/multus-k9pcc" podUID="4181bfb8-478b-4b30-a2cb-744da1010aeb" Nov 26 06:27:13 crc kubenswrapper[4775]: I1126 06:27:13.033311 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/1.log" Nov 26 06:27:13 crc kubenswrapper[4775]: I1126 06:27:13.327412 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:13 crc kubenswrapper[4775]: I1126 06:27:13.327463 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:13 crc kubenswrapper[4775]: I1126 06:27:13.327414 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:13 crc kubenswrapper[4775]: E1126 06:27:13.327578 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:13 crc kubenswrapper[4775]: I1126 06:27:13.327653 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:13 crc kubenswrapper[4775]: E1126 06:27:13.327691 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:13 crc kubenswrapper[4775]: E1126 06:27:13.327937 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:13 crc kubenswrapper[4775]: E1126 06:27:13.328006 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:15 crc kubenswrapper[4775]: I1126 06:27:15.326996 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:15 crc kubenswrapper[4775]: E1126 06:27:15.327967 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:15 crc kubenswrapper[4775]: I1126 06:27:15.327131 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:15 crc kubenswrapper[4775]: I1126 06:27:15.327165 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:15 crc kubenswrapper[4775]: I1126 06:27:15.327133 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:15 crc kubenswrapper[4775]: E1126 06:27:15.328261 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:15 crc kubenswrapper[4775]: E1126 06:27:15.328418 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:15 crc kubenswrapper[4775]: E1126 06:27:15.328561 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:17 crc kubenswrapper[4775]: E1126 06:27:17.305165 4775 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 26 06:27:17 crc kubenswrapper[4775]: I1126 06:27:17.326992 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:17 crc kubenswrapper[4775]: E1126 06:27:17.327209 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:17 crc kubenswrapper[4775]: I1126 06:27:17.327246 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:17 crc kubenswrapper[4775]: I1126 06:27:17.327307 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:17 crc kubenswrapper[4775]: I1126 06:27:17.329502 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:17 crc kubenswrapper[4775]: E1126 06:27:17.329505 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:17 crc kubenswrapper[4775]: E1126 06:27:17.329625 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:17 crc kubenswrapper[4775]: E1126 06:27:17.329768 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:17 crc kubenswrapper[4775]: E1126 06:27:17.448671 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 06:27:19 crc kubenswrapper[4775]: I1126 06:27:19.327698 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:19 crc kubenswrapper[4775]: I1126 06:27:19.327853 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:19 crc kubenswrapper[4775]: I1126 06:27:19.327771 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:19 crc kubenswrapper[4775]: E1126 06:27:19.328013 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:19 crc kubenswrapper[4775]: I1126 06:27:19.328059 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:19 crc kubenswrapper[4775]: E1126 06:27:19.328225 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:19 crc kubenswrapper[4775]: E1126 06:27:19.328355 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:19 crc kubenswrapper[4775]: E1126 06:27:19.328454 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:19 crc kubenswrapper[4775]: I1126 06:27:19.329585 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:27:20 crc kubenswrapper[4775]: I1126 06:27:20.062094 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/3.log" Nov 26 06:27:20 crc kubenswrapper[4775]: I1126 06:27:20.065188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerStarted","Data":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} Nov 26 06:27:20 crc kubenswrapper[4775]: I1126 06:27:20.065618 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:27:20 crc kubenswrapper[4775]: I1126 06:27:20.100369 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podStartSLOduration=103.100350591 podStartE2EDuration="1m43.100350591s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:20.095764253 +0000 UTC m=+123.457068225" watchObservedRunningTime="2025-11-26 06:27:20.100350591 +0000 UTC m=+123.461654553" Nov 26 06:27:20 crc kubenswrapper[4775]: I1126 06:27:20.287755 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n98r8"] Nov 26 06:27:20 crc kubenswrapper[4775]: I1126 06:27:20.287955 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:20 crc kubenswrapper[4775]: E1126 06:27:20.288130 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:21 crc kubenswrapper[4775]: I1126 06:27:21.327411 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:21 crc kubenswrapper[4775]: I1126 06:27:21.327434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:21 crc kubenswrapper[4775]: E1126 06:27:21.327914 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:21 crc kubenswrapper[4775]: I1126 06:27:21.327356 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:21 crc kubenswrapper[4775]: I1126 06:27:21.327440 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:21 crc kubenswrapper[4775]: E1126 06:27:21.328205 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:21 crc kubenswrapper[4775]: E1126 06:27:21.328334 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:21 crc kubenswrapper[4775]: E1126 06:27:21.328565 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:22 crc kubenswrapper[4775]: E1126 06:27:22.449668 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 06:27:23 crc kubenswrapper[4775]: I1126 06:27:23.327120 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:23 crc kubenswrapper[4775]: E1126 06:27:23.327309 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:23 crc kubenswrapper[4775]: I1126 06:27:23.327589 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:23 crc kubenswrapper[4775]: E1126 06:27:23.327701 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:23 crc kubenswrapper[4775]: I1126 06:27:23.327930 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:23 crc kubenswrapper[4775]: E1126 06:27:23.328027 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:23 crc kubenswrapper[4775]: I1126 06:27:23.328138 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:23 crc kubenswrapper[4775]: E1126 06:27:23.328437 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:25 crc kubenswrapper[4775]: I1126 06:27:25.327555 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:25 crc kubenswrapper[4775]: I1126 06:27:25.327614 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:25 crc kubenswrapper[4775]: E1126 06:27:25.327805 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:25 crc kubenswrapper[4775]: I1126 06:27:25.327857 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:25 crc kubenswrapper[4775]: I1126 06:27:25.327870 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:25 crc kubenswrapper[4775]: E1126 06:27:25.327988 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:25 crc kubenswrapper[4775]: E1126 06:27:25.328104 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:25 crc kubenswrapper[4775]: E1126 06:27:25.328186 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:26 crc kubenswrapper[4775]: I1126 06:27:26.327339 4775 scope.go:117] "RemoveContainer" containerID="a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94" Nov 26 06:27:26 crc kubenswrapper[4775]: I1126 06:27:26.509085 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:27:27 crc kubenswrapper[4775]: I1126 06:27:27.092984 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/1.log" Nov 26 06:27:27 crc kubenswrapper[4775]: I1126 06:27:27.093065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerStarted","Data":"8dd11dfd36979f712eecdab186571b6768bcbde711a197ba3cd3465503d292ee"} Nov 26 06:27:27 crc kubenswrapper[4775]: I1126 06:27:27.327486 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:27 crc kubenswrapper[4775]: I1126 06:27:27.327568 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:27 crc kubenswrapper[4775]: E1126 06:27:27.329374 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:27 crc kubenswrapper[4775]: I1126 06:27:27.329448 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:27 crc kubenswrapper[4775]: I1126 06:27:27.329459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:27 crc kubenswrapper[4775]: E1126 06:27:27.329589 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:27 crc kubenswrapper[4775]: E1126 06:27:27.329827 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:27 crc kubenswrapper[4775]: E1126 06:27:27.330000 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:27 crc kubenswrapper[4775]: E1126 06:27:27.450488 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 06:27:29 crc kubenswrapper[4775]: I1126 06:27:29.327706 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:29 crc kubenswrapper[4775]: I1126 06:27:29.327817 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:29 crc kubenswrapper[4775]: I1126 06:27:29.327860 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:29 crc kubenswrapper[4775]: E1126 06:27:29.328477 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:29 crc kubenswrapper[4775]: E1126 06:27:29.328284 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:29 crc kubenswrapper[4775]: I1126 06:27:29.327933 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:29 crc kubenswrapper[4775]: E1126 06:27:29.328564 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:29 crc kubenswrapper[4775]: E1126 06:27:29.328635 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:31 crc kubenswrapper[4775]: I1126 06:27:31.326644 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:31 crc kubenswrapper[4775]: I1126 06:27:31.326766 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:31 crc kubenswrapper[4775]: E1126 06:27:31.326879 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 06:27:31 crc kubenswrapper[4775]: I1126 06:27:31.326919 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:31 crc kubenswrapper[4775]: I1126 06:27:31.327025 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:31 crc kubenswrapper[4775]: E1126 06:27:31.327188 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n98r8" podUID="1e24b0f5-8555-40a4-a965-e67d5cd305a6" Nov 26 06:27:31 crc kubenswrapper[4775]: E1126 06:27:31.327354 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 06:27:31 crc kubenswrapper[4775]: E1126 06:27:31.327476 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.327353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.327368 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.327662 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.327686 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.331522 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.333612 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.334002 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.334437 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.334571 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 26 06:27:33 crc kubenswrapper[4775]: I1126 06:27:33.336232 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.937932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.986191 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts"] Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.986888 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.987434 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p4qsm"] Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.988318 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.988876 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs"] Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.989556 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kclct"] Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.989598 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.989877 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.991201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.994317 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.994346 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.994371 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.994408 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.995217 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.995296 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.996044 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm"] Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.998643 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.998863 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.998951 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.999257 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.999451 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 26 06:27:35 crc kubenswrapper[4775]: I1126 06:27:35.999911 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.005377 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.005746 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-dqwb7"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.006204 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.024901 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.025310 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.026261 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.026282 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.027004 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.028598 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.029015 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.029341 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.029602 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.029881 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.030091 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.030323 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031160 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-44mw9"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.032101 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031257 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031260 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031392 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031447 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031638 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031710 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031791 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.044610 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031866 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.031926 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.038516 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.045279 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.039263 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.039327 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.045511 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.045588 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.046420 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.046948 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.047681 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.060430 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.060763 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061349 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrf87\" (UniqueName: \"kubernetes.io/projected/e0344f36-af1e-44ad-9f72-7a616d6e222c-kube-api-access-zrf87\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061399 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-client-ca\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-config\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061467 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d05b88a6-499b-4c39-bb86-5629116aa323-serving-cert\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061496 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-etcd-client\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061523 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ed720241-da26-4f88-9838-c851d15fd2d8-audit-dir\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061579 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdrwx\" (UniqueName: \"kubernetes.io/projected/ab5f6914-f92e-47d6-9810-b61ce5151538-kube-api-access-sdrwx\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061609 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-encryption-config\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061637 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zztgx\" (UniqueName: \"kubernetes.io/projected/4942219a-29bb-4f6b-904e-e54d0222d948-kube-api-access-zztgx\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061665 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-config\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061691 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-etcd-client\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-serving-cert\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061787 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4942219a-29bb-4f6b-904e-e54d0222d948-machine-approver-tls\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061819 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxljn\" (UniqueName: \"kubernetes.io/projected/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-kube-api-access-gxljn\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-serving-cert\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061918 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zprvn\" (UniqueName: \"kubernetes.io/projected/2278df09-65b1-440e-a613-7bea0b8c73cd-kube-api-access-zprvn\") pod \"downloads-7954f5f757-44mw9\" (UID: \"2278df09-65b1-440e-a613-7bea0b8c73cd\") " pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4942219a-29bb-4f6b-904e-e54d0222d948-config\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.061974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e0344f36-af1e-44ad-9f72-7a616d6e222c-node-pullsecrets\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062006 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-config\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-config\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82ht6\" (UniqueName: \"kubernetes.io/projected/d05b88a6-499b-4c39-bb86-5629116aa323-kube-api-access-82ht6\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-audit-policies\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062132 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsb46\" (UniqueName: \"kubernetes.io/projected/ed720241-da26-4f88-9838-c851d15fd2d8-kube-api-access-tsb46\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-client-ca\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062190 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4942219a-29bb-4f6b-904e-e54d0222d948-auth-proxy-config\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062261 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0344f36-af1e-44ad-9f72-7a616d6e222c-audit-dir\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062295 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-image-import-ca\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-audit\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062426 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-images\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062460 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5f6914-f92e-47d6-9810-b61ce5151538-serving-cert\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062493 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-encryption-config\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062521 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.062924 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vl6j8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.063322 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.063360 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.063906 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.064482 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-49tww"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.064736 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.066738 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.067031 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.067132 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4bw62"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.067446 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.067582 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.067773 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.067458 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.070386 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.071021 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.071662 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.072549 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.073163 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.074210 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.074929 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.075995 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.076654 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.077585 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.078097 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.078615 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.079022 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.080446 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-drckl"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.080935 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.081595 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.082319 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.090492 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4h8zh"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091070 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cv6pw"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091340 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091493 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091628 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091820 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091942 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.091983 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.092281 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.092387 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.092562 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.092663 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.092674 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.093164 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.094885 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.095955 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.095983 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096122 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096138 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096245 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096307 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096384 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096444 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096518 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096575 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096663 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096768 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096903 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.096955 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097040 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097064 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097117 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097146 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097190 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097236 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097147 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097037 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097401 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097438 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097461 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097486 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097529 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097566 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097597 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097486 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097636 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097705 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.097782 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.098286 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.112429 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.115656 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.115702 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.116148 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.116449 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.116457 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.140561 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.141351 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.141701 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.141892 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.142089 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.142560 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.143233 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zbzr7"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.143851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.145578 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.145706 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.148576 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.149064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.150070 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-szmvm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.150853 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.151912 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.152140 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.152480 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.155690 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.155846 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.157524 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.162425 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rtsmv"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.162995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtbj2\" (UniqueName: \"kubernetes.io/projected/268d85ad-f165-4cd2-931f-1da2caefe3cd-kube-api-access-wtbj2\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163027 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163051 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-serving-cert\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zprvn\" (UniqueName: \"kubernetes.io/projected/2278df09-65b1-440e-a613-7bea0b8c73cd-kube-api-access-zprvn\") pod \"downloads-7954f5f757-44mw9\" (UID: \"2278df09-65b1-440e-a613-7bea0b8c73cd\") " pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163089 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4942219a-29bb-4f6b-904e-e54d0222d948-config\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163106 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-serving-cert\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163122 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvvd9\" (UniqueName: \"kubernetes.io/projected/b188fc55-2f3e-40a2-a779-863cacc2a1eb-kube-api-access-mvvd9\") pod \"cluster-samples-operator-665b6dd947-8lbfh\" (UID: \"b188fc55-2f3e-40a2-a779-863cacc2a1eb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163138 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61d701be-8e8d-4d60-a562-328010fc2003-audit-dir\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163168 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtdqd\" (UniqueName: \"kubernetes.io/projected/61d701be-8e8d-4d60-a562-328010fc2003-kube-api-access-xtdqd\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82ht6\" (UniqueName: \"kubernetes.io/projected/d05b88a6-499b-4c39-bb86-5629116aa323-kube-api-access-82ht6\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e0344f36-af1e-44ad-9f72-7a616d6e222c-node-pullsecrets\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163223 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-config\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-config\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-audit-policies\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163267 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-serving-cert\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163285 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163302 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vjzp\" (UniqueName: \"kubernetes.io/projected/23e94d9b-5940-42b3-9876-17fc7bf7fa18-kube-api-access-5vjzp\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163319 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsb46\" (UniqueName: \"kubernetes.io/projected/ed720241-da26-4f88-9838-c851d15fd2d8-kube-api-access-tsb46\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163335 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xbxz\" (UniqueName: \"kubernetes.io/projected/9a2a9335-a9d7-4145-af84-2084ad5d8c14-kube-api-access-4xbxz\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163351 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w927d\" (UniqueName: \"kubernetes.io/projected/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-kube-api-access-w927d\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163386 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-client-ca\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-stats-auth\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163422 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163446 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163462 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0344f36-af1e-44ad-9f72-7a616d6e222c-audit-dir\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163477 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4942219a-29bb-4f6b-904e-e54d0222d948-auth-proxy-config\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163492 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/283bf004-b2c0-4c8c-b979-a884cf514ab2-proxy-tls\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163509 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-image-import-ca\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163524 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-service-ca-bundle\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163557 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163584 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163604 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-audit\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163625 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163769 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2a9335-a9d7-4145-af84-2084ad5d8c14-serving-cert\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-default-certificate\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163802 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-images\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163820 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5f6914-f92e-47d6-9810-b61ce5151538-serving-cert\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163844 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-encryption-config\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163862 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163878 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-service-ca\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163899 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-metrics-certs\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163927 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163947 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-trusted-ca-bundle\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.163978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrf87\" (UniqueName: \"kubernetes.io/projected/e0344f36-af1e-44ad-9f72-7a616d6e222c-kube-api-access-zrf87\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164034 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23e94d9b-5940-42b3-9876-17fc7bf7fa18-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-client-ca\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164064 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-config\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164079 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/23e94d9b-5940-42b3-9876-17fc7bf7fa18-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164093 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/23e94d9b-5940-42b3-9876-17fc7bf7fa18-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-config\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164129 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/283bf004-b2c0-4c8c-b979-a884cf514ab2-images\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9a2a9335-a9d7-4145-af84-2084ad5d8c14-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164165 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-audit-policies\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d05b88a6-499b-4c39-bb86-5629116aa323-serving-cert\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164209 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-etcd-client\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164228 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ed720241-da26-4f88-9838-c851d15fd2d8-audit-dir\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164315 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-oauth-config\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85tg8\" (UniqueName: \"kubernetes.io/projected/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-kube-api-access-85tg8\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164360 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdrwx\" (UniqueName: \"kubernetes.io/projected/ab5f6914-f92e-47d6-9810-b61ce5151538-kube-api-access-sdrwx\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164374 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-encryption-config\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164391 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zztgx\" (UniqueName: \"kubernetes.io/projected/4942219a-29bb-4f6b-904e-e54d0222d948-kube-api-access-zztgx\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164406 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-config\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164423 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-etcd-client\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b188fc55-2f3e-40a2-a779-863cacc2a1eb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8lbfh\" (UID: \"b188fc55-2f3e-40a2-a779-863cacc2a1eb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164455 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-serving-cert\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-service-ca-bundle\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164511 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4942219a-29bb-4f6b-904e-e54d0222d948-machine-approver-tls\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164526 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164543 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/283bf004-b2c0-4c8c-b979-a884cf514ab2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5s6k\" (UniqueName: \"kubernetes.io/projected/283bf004-b2c0-4c8c-b979-a884cf514ab2-kube-api-access-q5s6k\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164571 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-oauth-serving-cert\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxljn\" (UniqueName: \"kubernetes.io/projected/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-kube-api-access-gxljn\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.164602 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-config\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.165677 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-client-ca\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.165885 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e0344f36-af1e-44ad-9f72-7a616d6e222c-audit-dir\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.166505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4942219a-29bb-4f6b-904e-e54d0222d948-auth-proxy-config\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.166661 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.168023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-image-import-ca\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.168084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4942219a-29bb-4f6b-904e-e54d0222d948-config\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.168515 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n4dwr"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.168869 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.169169 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.169473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.169809 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.170878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.170906 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.171590 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-config\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.171642 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-config\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.171772 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.171805 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e0344f36-af1e-44ad-9f72-7a616d6e222c-node-pullsecrets\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.172357 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-config\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.172453 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ed720241-da26-4f88-9838-c851d15fd2d8-audit-policies\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.172458 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-audit\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.173366 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.173842 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-etcd-serving-ca\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.175153 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-etcd-client\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.175287 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4942219a-29bb-4f6b-904e-e54d0222d948-machine-approver-tls\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.175481 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ed720241-da26-4f88-9838-c851d15fd2d8-audit-dir\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.176127 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-client-ca\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.176304 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-config\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.176605 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.176834 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0344f36-af1e-44ad-9f72-7a616d6e222c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.177018 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-images\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.177040 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-encryption-config\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.178105 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-serving-cert\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.178483 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed720241-da26-4f88-9838-c851d15fd2d8-serving-cert\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.178525 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-encryption-config\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.179147 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d05b88a6-499b-4c39-bb86-5629116aa323-serving-cert\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.179690 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.179999 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5f6914-f92e-47d6-9810-b61ce5151538-serving-cert\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.180203 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z59ck"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.180304 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.180427 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.181689 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.181780 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e0344f36-af1e-44ad-9f72-7a616d6e222c-etcd-client\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.181932 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.182088 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.182678 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gt6r8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.183150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kclct"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.183166 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pv8cb"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.183738 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.184207 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.184791 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.188894 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p4qsm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.188920 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-dqwb7"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.188932 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.188941 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-49tww"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.189019 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.191856 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cv6pw"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.196298 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.199064 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.203462 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.208892 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.210302 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4bw62"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.210424 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.213762 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.214998 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.216350 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vl6j8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.217329 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.218384 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zbzr7"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.219386 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.220816 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-44mw9"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.223954 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.224125 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.224614 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4h8zh"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.226311 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.227586 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.228510 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.229167 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.230163 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.231406 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.233274 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.236217 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-szmvm"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.237722 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rp4jq"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.238234 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.239495 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-9cq9z"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.240023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.240673 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gt6r8"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.242247 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.244071 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n4dwr"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.245448 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.246519 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z59ck"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.247681 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.248104 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.249077 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rtsmv"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.250121 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pv8cb"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.251368 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.252238 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9cq9z"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.253270 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rp4jq"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.254305 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lt2r2"] Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.255602 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265304 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-service-ca\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265363 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-metrics-certs\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265426 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-trusted-ca-bundle\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265511 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23e94d9b-5940-42b3-9876-17fc7bf7fa18-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265568 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-config\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/23e94d9b-5940-42b3-9876-17fc7bf7fa18-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265610 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/283bf004-b2c0-4c8c-b979-a884cf514ab2-images\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265630 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9a2a9335-a9d7-4145-af84-2084ad5d8c14-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/23e94d9b-5940-42b3-9876-17fc7bf7fa18-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265675 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-audit-policies\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265704 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-oauth-config\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85tg8\" (UniqueName: \"kubernetes.io/projected/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-kube-api-access-85tg8\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265804 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b188fc55-2f3e-40a2-a779-863cacc2a1eb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8lbfh\" (UID: \"b188fc55-2f3e-40a2-a779-863cacc2a1eb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265878 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-service-ca-bundle\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/283bf004-b2c0-4c8c-b979-a884cf514ab2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265952 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5s6k\" (UniqueName: \"kubernetes.io/projected/283bf004-b2c0-4c8c-b979-a884cf514ab2-kube-api-access-q5s6k\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.265975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-oauth-serving-cert\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266012 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-config\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266047 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtbj2\" (UniqueName: \"kubernetes.io/projected/268d85ad-f165-4cd2-931f-1da2caefe3cd-kube-api-access-wtbj2\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266102 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61d701be-8e8d-4d60-a562-328010fc2003-audit-dir\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtdqd\" (UniqueName: \"kubernetes.io/projected/61d701be-8e8d-4d60-a562-328010fc2003-kube-api-access-xtdqd\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-serving-cert\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvvd9\" (UniqueName: \"kubernetes.io/projected/b188fc55-2f3e-40a2-a779-863cacc2a1eb-kube-api-access-mvvd9\") pod \"cluster-samples-operator-665b6dd947-8lbfh\" (UID: \"b188fc55-2f3e-40a2-a779-863cacc2a1eb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266228 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-serving-cert\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266235 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-service-ca\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vjzp\" (UniqueName: \"kubernetes.io/projected/23e94d9b-5940-42b3-9876-17fc7bf7fa18-kube-api-access-5vjzp\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266321 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xbxz\" (UniqueName: \"kubernetes.io/projected/9a2a9335-a9d7-4145-af84-2084ad5d8c14-kube-api-access-4xbxz\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w927d\" (UniqueName: \"kubernetes.io/projected/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-kube-api-access-w927d\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266356 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-stats-auth\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266403 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266426 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/283bf004-b2c0-4c8c-b979-a884cf514ab2-proxy-tls\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266466 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266485 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-service-ca-bundle\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266508 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266523 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-default-certificate\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266562 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2a9335-a9d7-4145-af84-2084ad5d8c14-serving-cert\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.266748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-trusted-ca-bundle\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.267188 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9a2a9335-a9d7-4145-af84-2084ad5d8c14-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.267255 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61d701be-8e8d-4d60-a562-328010fc2003-audit-dir\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.268028 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/283bf004-b2c0-4c8c-b979-a884cf514ab2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.268225 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-metrics-certs\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.268920 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-config\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.268941 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-oauth-serving-cert\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.268972 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/23e94d9b-5940-42b3-9876-17fc7bf7fa18-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.270122 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-config\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.270165 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-service-ca-bundle\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.270576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/23e94d9b-5940-42b3-9876-17fc7bf7fa18-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.270979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a2a9335-a9d7-4145-af84-2084ad5d8c14-serving-cert\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.270984 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/283bf004-b2c0-4c8c-b979-a884cf514ab2-images\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.271456 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.271784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-oauth-config\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.272063 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.272223 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-serving-cert\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.272331 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-stats-auth\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.272414 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-default-certificate\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.273227 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.273622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-service-ca-bundle\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.274457 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b188fc55-2f3e-40a2-a779-863cacc2a1eb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8lbfh\" (UID: \"b188fc55-2f3e-40a2-a779-863cacc2a1eb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.274831 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-serving-cert\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.274958 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/283bf004-b2c0-4c8c-b979-a884cf514ab2-proxy-tls\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.281335 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.289905 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.301919 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.308854 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.317547 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.328459 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.348806 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.360618 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.386858 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.392529 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.394867 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.403281 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.409105 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.420694 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.429352 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.449044 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.449788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-audit-policies\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.469734 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.478097 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.489309 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.498260 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.516499 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.527176 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.528752 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.548755 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.570070 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.588665 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.609205 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.629475 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.649214 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.670074 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.709223 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.729048 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.769000 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.790567 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.809852 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.829263 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.850114 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.869118 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.891370 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.909475 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.930140 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.951198 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.970029 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 26 06:27:36 crc kubenswrapper[4775]: I1126 06:27:36.990343 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.009577 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.056147 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsb46\" (UniqueName: \"kubernetes.io/projected/ed720241-da26-4f88-9838-c851d15fd2d8-kube-api-access-tsb46\") pod \"apiserver-7bbb656c7d-kn6ts\" (UID: \"ed720241-da26-4f88-9838-c851d15fd2d8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.076934 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zprvn\" (UniqueName: \"kubernetes.io/projected/2278df09-65b1-440e-a613-7bea0b8c73cd-kube-api-access-zprvn\") pod \"downloads-7954f5f757-44mw9\" (UID: \"2278df09-65b1-440e-a613-7bea0b8c73cd\") " pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.089744 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.093961 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zztgx\" (UniqueName: \"kubernetes.io/projected/4942219a-29bb-4f6b-904e-e54d0222d948-kube-api-access-zztgx\") pod \"machine-approver-56656f9798-26sfm\" (UID: \"4942219a-29bb-4f6b-904e-e54d0222d948\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.109834 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.129430 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.151230 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.169831 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.187532 4775 request.go:700] Waited for 1.016198086s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/configmaps?fieldSelector=metadata.name%3Detcd-ca-bundle&limit=500&resourceVersion=0 Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.189586 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.210378 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.215119 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.229442 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.279104 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxljn\" (UniqueName: \"kubernetes.io/projected/74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb-kube-api-access-gxljn\") pod \"machine-api-operator-5694c8668f-p4qsm\" (UID: \"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.290179 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.292021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82ht6\" (UniqueName: \"kubernetes.io/projected/d05b88a6-499b-4c39-bb86-5629116aa323-kube-api-access-82ht6\") pod \"route-controller-manager-6576b87f9c-84sgs\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.312762 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.330664 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.355338 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.355867 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.367945 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.370161 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: W1126 06:27:37.399877 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4942219a_29bb_4f6b_904e_e54d0222d948.slice/crio-9ef6d2a0b6c72837a50fd7d36f073d25ed6e168bddfb6908068a9a705ac65786 WatchSource:0}: Error finding container 9ef6d2a0b6c72837a50fd7d36f073d25ed6e168bddfb6908068a9a705ac65786: Status 404 returned error can't find the container with id 9ef6d2a0b6c72837a50fd7d36f073d25ed6e168bddfb6908068a9a705ac65786 Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.419366 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrf87\" (UniqueName: \"kubernetes.io/projected/e0344f36-af1e-44ad-9f72-7a616d6e222c-kube-api-access-zrf87\") pod \"apiserver-76f77b778f-kclct\" (UID: \"e0344f36-af1e-44ad-9f72-7a616d6e222c\") " pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.432483 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.443700 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdrwx\" (UniqueName: \"kubernetes.io/projected/ab5f6914-f92e-47d6-9810-b61ce5151538-kube-api-access-sdrwx\") pod \"controller-manager-879f6c89f-dqwb7\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.449489 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.469604 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.479012 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts"] Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.489853 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 26 06:27:37 crc kubenswrapper[4775]: W1126 06:27:37.501054 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded720241_da26_4f88_9838_c851d15fd2d8.slice/crio-29059e38acb293e1f3de9e4a97918c1ab928b13e762e07f93f979e07b60a976f WatchSource:0}: Error finding container 29059e38acb293e1f3de9e4a97918c1ab928b13e762e07f93f979e07b60a976f: Status 404 returned error can't find the container with id 29059e38acb293e1f3de9e4a97918c1ab928b13e762e07f93f979e07b60a976f Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.509676 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.530165 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.533914 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.549945 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.568133 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.578702 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.579750 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-44mw9"] Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.589843 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.591150 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.610007 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.615293 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.630140 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.656584 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.673902 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.690936 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.708597 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.734093 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.753194 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.754092 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-p4qsm"] Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.769603 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.789012 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.795652 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs"] Nov 26 06:27:37 crc kubenswrapper[4775]: W1126 06:27:37.806003 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd05b88a6_499b_4c39_bb86_5629116aa323.slice/crio-cd7bb8b7d387818d91504eac98d002cb4e889e53e07b00839efcd65c2b995a8d WatchSource:0}: Error finding container cd7bb8b7d387818d91504eac98d002cb4e889e53e07b00839efcd65c2b995a8d: Status 404 returned error can't find the container with id cd7bb8b7d387818d91504eac98d002cb4e889e53e07b00839efcd65c2b995a8d Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.808314 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.817523 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kclct"] Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.829030 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: W1126 06:27:37.830112 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0344f36_af1e_44ad_9f72_7a616d6e222c.slice/crio-b1ca42efb94c4043fadf8660c6f878ea616299033852f44d97b0e26c4b090536 WatchSource:0}: Error finding container b1ca42efb94c4043fadf8660c6f878ea616299033852f44d97b0e26c4b090536: Status 404 returned error can't find the container with id b1ca42efb94c4043fadf8660c6f878ea616299033852f44d97b0e26c4b090536 Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.864706 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.871111 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.895521 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.909038 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.929910 4775 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.948772 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.969197 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 26 06:27:37 crc kubenswrapper[4775]: I1126 06:27:37.990085 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.009783 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.029701 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.087559 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.092607 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.093381 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.110138 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.119245 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-dqwb7"] Nov 26 06:27:38 crc kubenswrapper[4775]: W1126 06:27:38.127739 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab5f6914_f92e_47d6_9810_b61ce5151538.slice/crio-a59a9a6766290842991eb3a0047b312e59a57531830cb960d435ba83d4285ccc WatchSource:0}: Error finding container a59a9a6766290842991eb3a0047b312e59a57531830cb960d435ba83d4285ccc: Status 404 returned error can't find the container with id a59a9a6766290842991eb3a0047b312e59a57531830cb960d435ba83d4285ccc Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.130803 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.149645 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.162501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kclct" event={"ID":"e0344f36-af1e-44ad-9f72-7a616d6e222c","Type":"ContainerStarted","Data":"b1ca42efb94c4043fadf8660c6f878ea616299033852f44d97b0e26c4b090536"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.165287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" event={"ID":"d05b88a6-499b-4c39-bb86-5629116aa323","Type":"ContainerStarted","Data":"45f1d91905824ec0182e472333c57e9ea38390475ca09e3c4007a501545c4ef5"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.165311 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" event={"ID":"d05b88a6-499b-4c39-bb86-5629116aa323","Type":"ContainerStarted","Data":"cd7bb8b7d387818d91504eac98d002cb4e889e53e07b00839efcd65c2b995a8d"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.166029 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.167392 4775 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-84sgs container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.168944 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" podUID="d05b88a6-499b-4c39-bb86-5629116aa323" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.169000 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.169505 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" event={"ID":"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb","Type":"ContainerStarted","Data":"de1bdcbfe09485e8bab57590d391af336ad5e258af5e9def7dcf54d6f1591bba"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.169534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" event={"ID":"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb","Type":"ContainerStarted","Data":"fa94efd6d2f3d10b5b59db857bc10950fce03a2c365eec77d9a8bc51065bc981"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.171232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" event={"ID":"ab5f6914-f92e-47d6-9810-b61ce5151538","Type":"ContainerStarted","Data":"a59a9a6766290842991eb3a0047b312e59a57531830cb960d435ba83d4285ccc"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.172387 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-44mw9" event={"ID":"2278df09-65b1-440e-a613-7bea0b8c73cd","Type":"ContainerStarted","Data":"9294ae8ee0268cbcbd51de220723a7d48bd3c98c2277f876b1eb4dafbfb0d453"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.172412 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-44mw9" event={"ID":"2278df09-65b1-440e-a613-7bea0b8c73cd","Type":"ContainerStarted","Data":"946c4a29e98e8b8936fc64a4d5cafa648cca46954abceba5718b46589909c96a"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.172979 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.173684 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-44mw9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.173776 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-44mw9" podUID="2278df09-65b1-440e-a613-7bea0b8c73cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.175024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" event={"ID":"4942219a-29bb-4f6b-904e-e54d0222d948","Type":"ContainerStarted","Data":"f35b8fc35fc0ca2d66da67bef4660d5fd9a6c907149db2a0b64f61b0cd596ada"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.175052 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" event={"ID":"4942219a-29bb-4f6b-904e-e54d0222d948","Type":"ContainerStarted","Data":"9ef6d2a0b6c72837a50fd7d36f073d25ed6e168bddfb6908068a9a705ac65786"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.177908 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" event={"ID":"ed720241-da26-4f88-9838-c851d15fd2d8","Type":"ContainerDied","Data":"29c0b3e978d59f96db2746b1465ce6fc1f9e2059d10873d75dda5c014d6a8074"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.177805 4775 generic.go:334] "Generic (PLEG): container finished" podID="ed720241-da26-4f88-9838-c851d15fd2d8" containerID="29c0b3e978d59f96db2746b1465ce6fc1f9e2059d10873d75dda5c014d6a8074" exitCode=0 Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.178376 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" event={"ID":"ed720241-da26-4f88-9838-c851d15fd2d8","Type":"ContainerStarted","Data":"29059e38acb293e1f3de9e4a97918c1ab928b13e762e07f93f979e07b60a976f"} Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.188162 4775 request.go:700] Waited for 1.921033731s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.205202 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtbj2\" (UniqueName: \"kubernetes.io/projected/268d85ad-f165-4cd2-931f-1da2caefe3cd-kube-api-access-wtbj2\") pod \"console-f9d7485db-vl6j8\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.222598 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtdqd\" (UniqueName: \"kubernetes.io/projected/61d701be-8e8d-4d60-a562-328010fc2003-kube-api-access-xtdqd\") pod \"oauth-openshift-558db77b4-cv6pw\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.246059 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vjzp\" (UniqueName: \"kubernetes.io/projected/23e94d9b-5940-42b3-9876-17fc7bf7fa18-kube-api-access-5vjzp\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.266752 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xbxz\" (UniqueName: \"kubernetes.io/projected/9a2a9335-a9d7-4145-af84-2084ad5d8c14-kube-api-access-4xbxz\") pod \"openshift-config-operator-7777fb866f-8dhs8\" (UID: \"9a2a9335-a9d7-4145-af84-2084ad5d8c14\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.276845 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.282581 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w927d\" (UniqueName: \"kubernetes.io/projected/b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b-kube-api-access-w927d\") pod \"authentication-operator-69f744f599-4bw62\" (UID: \"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.303058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5s6k\" (UniqueName: \"kubernetes.io/projected/283bf004-b2c0-4c8c-b979-a884cf514ab2-kube-api-access-q5s6k\") pod \"machine-config-operator-74547568cd-vs44x\" (UID: \"283bf004-b2c0-4c8c-b979-a884cf514ab2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.327368 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/23e94d9b-5940-42b3-9876-17fc7bf7fa18-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fxkqj\" (UID: \"23e94d9b-5940-42b3-9876-17fc7bf7fa18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.342316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.344967 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.346643 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85tg8\" (UniqueName: \"kubernetes.io/projected/f02e8337-4b5c-4ec3-b0a2-38d674cbfe54-kube-api-access-85tg8\") pod \"router-default-5444994796-drckl\" (UID: \"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54\") " pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.365349 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvvd9\" (UniqueName: \"kubernetes.io/projected/b188fc55-2f3e-40a2-a779-863cacc2a1eb-kube-api-access-mvvd9\") pod \"cluster-samples-operator-665b6dd947-8lbfh\" (UID: \"b188fc55-2f3e-40a2-a779-863cacc2a1eb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396075 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/088e2d8e-dfb4-4a5d-b681-d7d46573b629-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396128 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/784c7366-4335-438f-9cc7-6ed4580fc0a5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cskh\" (UniqueName: \"kubernetes.io/projected/e1cb46aa-dd22-49d0-9a14-dab875c02723-kube-api-access-4cskh\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396178 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-certificates\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-trusted-ca\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396238 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396262 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsll5\" (UniqueName: \"kubernetes.io/projected/9f88d0e7-a6b2-469d-98c6-86747dc3c112-kube-api-access-dsll5\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396322 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f88d0e7-a6b2-469d-98c6-86747dc3c112-config\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396348 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-bound-sa-token\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396405 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396447 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396477 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f88d0e7-a6b2-469d-98c6-86747dc3c112-trusted-ca\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396499 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l578j\" (UniqueName: \"kubernetes.io/projected/3316a24f-1cce-4358-a8b4-03b40f9ceec0-kube-api-access-l578j\") pod \"migrator-59844c95c7-rb4zf\" (UID: \"3316a24f-1cce-4358-a8b4-03b40f9ceec0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396567 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/784c7366-4335-438f-9cc7-6ed4580fc0a5-config\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396600 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dd1a5521-45f9-4377-970e-4710f6749fa5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zbzr7\" (UID: \"dd1a5521-45f9-4377-970e-4710f6749fa5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396627 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/088e2d8e-dfb4-4a5d-b681-d7d46573b629-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/784c7366-4335-438f-9cc7-6ed4580fc0a5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcxd4\" (UniqueName: \"kubernetes.io/projected/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-kube-api-access-rcxd4\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-config\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396738 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f88d0e7-a6b2-469d-98c6-86747dc3c112-serving-cert\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396760 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8zrj\" (UniqueName: \"kubernetes.io/projected/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-kube-api-access-k8zrj\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396807 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-config\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1cb46aa-dd22-49d0-9a14-dab875c02723-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfhtm\" (UniqueName: \"kubernetes.io/projected/dd1a5521-45f9-4377-970e-4710f6749fa5-kube-api-access-lfhtm\") pod \"multus-admission-controller-857f4d67dd-zbzr7\" (UID: \"dd1a5521-45f9-4377-970e-4710f6749fa5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpfb8\" (UniqueName: \"kubernetes.io/projected/cb8fd937-d759-4e21-98af-5b833a8b3f52-kube-api-access-wpfb8\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5x92\" (UID: \"cb8fd937-d759-4e21-98af-5b833a8b3f52\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.396902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82hfb\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-kube-api-access-82hfb\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.397156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb8fd937-d759-4e21-98af-5b833a8b3f52-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5x92\" (UID: \"cb8fd937-d759-4e21-98af-5b833a8b3f52\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.397192 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-tls\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.397256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1cb46aa-dd22-49d0-9a14-dab875c02723-proxy-tls\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.397294 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.399187 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:38.899171423 +0000 UTC m=+142.260475455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.399431 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.405603 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.416155 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.418204 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498297 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.498583 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:38.998568127 +0000 UTC m=+142.359872079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498841 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eceaefdb-3bc4-4199-815f-c5cca21b8110-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498862 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-socket-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-plugins-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498892 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-config\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498910 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/088e2d8e-dfb4-4a5d-b681-d7d46573b629-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498936 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-trusted-ca\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498954 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjvnc\" (UniqueName: \"kubernetes.io/projected/2a863afc-708f-4afe-b3f4-c6cb950c23a2-kube-api-access-jjvnc\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498968 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-csi-data-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.498986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499001 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c83b476-9e46-440e-be8b-76e02c398f70-config\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a969da6-616f-4f99-8271-5a0d6a39b06c-metrics-tls\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-mountpoint-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfb25e56-d6ab-4e49-86de-9a20ba2ee02a-cert\") pod \"ingress-canary-rp4jq\" (UID: \"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a\") " pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499074 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499089 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szxmw\" (UniqueName: \"kubernetes.io/projected/8a969da6-616f-4f99-8271-5a0d6a39b06c-kube-api-access-szxmw\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499113 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-bound-sa-token\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499130 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499152 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499168 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-tmpfs\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2a863afc-708f-4afe-b3f4-c6cb950c23a2-node-bootstrap-token\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499210 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499246 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f88d0e7-a6b2-469d-98c6-86747dc3c112-trusted-ca\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499263 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l578j\" (UniqueName: \"kubernetes.io/projected/3316a24f-1cce-4358-a8b4-03b40f9ceec0-kube-api-access-l578j\") pod \"migrator-59844c95c7-rb4zf\" (UID: \"3316a24f-1cce-4358-a8b4-03b40f9ceec0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-apiservice-cert\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499330 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/784c7366-4335-438f-9cc7-6ed4580fc0a5-config\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499356 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dd1a5521-45f9-4377-970e-4710f6749fa5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zbzr7\" (UID: \"dd1a5521-45f9-4377-970e-4710f6749fa5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499371 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/088e2d8e-dfb4-4a5d-b681-d7d46573b629-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499387 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl5n7\" (UniqueName: \"kubernetes.io/projected/f1365993-a452-4e00-92e0-64e1c08ce1c2-kube-api-access-bl5n7\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499405 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b0fb0ce-6df9-4e4c-8d48-b35699664601-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vrt6m\" (UID: \"6b0fb0ce-6df9-4e4c-8d48-b35699664601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-config\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcxd4\" (UniqueName: \"kubernetes.io/projected/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-kube-api-access-rcxd4\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499488 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46jg9\" (UniqueName: \"kubernetes.io/projected/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-kube-api-access-46jg9\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499503 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-955ln\" (UniqueName: \"kubernetes.io/projected/6b0fb0ce-6df9-4e4c-8d48-b35699664601-kube-api-access-955ln\") pod \"package-server-manager-789f6589d5-vrt6m\" (UID: \"6b0fb0ce-6df9-4e4c-8d48-b35699664601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499520 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc93554a-637c-4af3-a239-bdfab76f780f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499540 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8zrj\" (UniqueName: \"kubernetes.io/projected/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-kube-api-access-k8zrj\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499573 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-config\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1cb46aa-dd22-49d0-9a14-dab875c02723-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499608 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc93554a-637c-4af3-a239-bdfab76f780f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499625 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/de5015a4-08fc-447e-86da-8341572f7143-srv-cert\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499642 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82hfb\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-kube-api-access-82hfb\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499658 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-config-volume\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkxhc\" (UniqueName: \"kubernetes.io/projected/cfb25e56-d6ab-4e49-86de-9a20ba2ee02a-kube-api-access-tkxhc\") pod \"ingress-canary-rp4jq\" (UID: \"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a\") " pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499692 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhgx9\" (UniqueName: \"kubernetes.io/projected/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-kube-api-access-dhgx9\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499707 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xglbk\" (UniqueName: \"kubernetes.io/projected/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-kube-api-access-xglbk\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499739 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-948ls\" (UniqueName: \"kubernetes.io/projected/f1c14a59-51f9-475d-8867-aefb952b66dc-kube-api-access-948ls\") pod \"dns-operator-744455d44c-gt6r8\" (UID: \"f1c14a59-51f9-475d-8867-aefb952b66dc\") " pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a969da6-616f-4f99-8271-5a0d6a39b06c-trusted-ca\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-tls\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499795 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc93554a-637c-4af3-a239-bdfab76f780f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499868 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1cb46aa-dd22-49d0-9a14-dab875c02723-proxy-tls\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499893 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kfs2\" (UniqueName: \"kubernetes.io/projected/54b5883f-4f48-4b9e-84b7-46eeb452aedc-kube-api-access-6kfs2\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499931 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/784c7366-4335-438f-9cc7-6ed4580fc0a5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499947 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cskh\" (UniqueName: \"kubernetes.io/projected/e1cb46aa-dd22-49d0-9a14-dab875c02723-kube-api-access-4cskh\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a11b6fbe-aabd-45e3-9302-3b3678bba644-metrics-tls\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.499987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-certificates\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500003 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wknk\" (UniqueName: \"kubernetes.io/projected/a11b6fbe-aabd-45e3-9302-3b3678bba644-kube-api-access-2wknk\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eceaefdb-3bc4-4199-815f-c5cca21b8110-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-webhook-cert\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500051 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsll5\" (UniqueName: \"kubernetes.io/projected/9f88d0e7-a6b2-469d-98c6-86747dc3c112-kube-api-access-dsll5\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500065 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-srv-cert\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500091 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a11b6fbe-aabd-45e3-9302-3b3678bba644-config-volume\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500107 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-signing-cabundle\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500123 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntf5k\" (UniqueName: \"kubernetes.io/projected/de5015a4-08fc-447e-86da-8341572f7143-kube-api-access-ntf5k\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f88d0e7-a6b2-469d-98c6-86747dc3c112-config\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500173 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-client\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500236 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54b5883f-4f48-4b9e-84b7-46eeb452aedc-serving-cert\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500253 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wns7g\" (UniqueName: \"kubernetes.io/projected/3c83b476-9e46-440e-be8b-76e02c398f70-kube-api-access-wns7g\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-secret-volume\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.500347 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501097 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a969da6-616f-4f99-8271-5a0d6a39b06c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501131 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/de5015a4-08fc-447e-86da-8341572f7143-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501169 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/784c7366-4335-438f-9cc7-6ed4580fc0a5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501210 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f88d0e7-a6b2-469d-98c6-86747dc3c112-serving-cert\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501230 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2a863afc-708f-4afe-b3f4-c6cb950c23a2-certs\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501274 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6c78\" (UniqueName: \"kubernetes.io/projected/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-kube-api-access-d6c78\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501305 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-ca\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfhtm\" (UniqueName: \"kubernetes.io/projected/dd1a5521-45f9-4377-970e-4710f6749fa5-kube-api-access-lfhtm\") pod \"multus-admission-controller-857f4d67dd-zbzr7\" (UID: \"dd1a5521-45f9-4377-970e-4710f6749fa5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpfb8\" (UniqueName: \"kubernetes.io/projected/cb8fd937-d759-4e21-98af-5b833a8b3f52-kube-api-access-wpfb8\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5x92\" (UID: \"cb8fd937-d759-4e21-98af-5b833a8b3f52\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501392 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb8fd937-d759-4e21-98af-5b833a8b3f52-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5x92\" (UID: \"cb8fd937-d759-4e21-98af-5b833a8b3f52\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501412 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-registration-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501463 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-signing-key\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501499 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-service-ca\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82rkd\" (UniqueName: \"kubernetes.io/projected/764b5255-c5f2-4e67-8ad9-0050044df158-kube-api-access-82rkd\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501816 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x78j7\" (UniqueName: \"kubernetes.io/projected/eceaefdb-3bc4-4199-815f-c5cca21b8110-kube-api-access-x78j7\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501871 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1c14a59-51f9-475d-8867-aefb952b66dc-metrics-tls\") pod \"dns-operator-744455d44c-gt6r8\" (UID: \"f1c14a59-51f9-475d-8867-aefb952b66dc\") " pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.501892 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c83b476-9e46-440e-be8b-76e02c398f70-serving-cert\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.502682 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-config\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.503468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f88d0e7-a6b2-469d-98c6-86747dc3c112-config\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.503603 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1cb46aa-dd22-49d0-9a14-dab875c02723-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.505447 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/088e2d8e-dfb4-4a5d-b681-d7d46573b629-ca-trust-extracted\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.506592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-certificates\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.506746 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/784c7366-4335-438f-9cc7-6ed4580fc0a5-config\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.506977 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-config\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.508191 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.008176572 +0000 UTC m=+142.369480524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.510393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/dd1a5521-45f9-4377-970e-4710f6749fa5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-zbzr7\" (UID: \"dd1a5521-45f9-4377-970e-4710f6749fa5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.510526 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.511110 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1cb46aa-dd22-49d0-9a14-dab875c02723-proxy-tls\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.511564 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-trusted-ca\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.516662 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f88d0e7-a6b2-469d-98c6-86747dc3c112-trusted-ca\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.516989 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.517103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f88d0e7-a6b2-469d-98c6-86747dc3c112-serving-cert\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.518611 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vl6j8"] Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.522996 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/784c7366-4335-438f-9cc7-6ed4580fc0a5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.523126 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cb8fd937-d759-4e21-98af-5b833a8b3f52-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5x92\" (UID: \"cb8fd937-d759-4e21-98af-5b833a8b3f52\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.523672 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.527027 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-tls\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.529335 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.538043 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/088e2d8e-dfb4-4a5d-b681-d7d46573b629-installation-pull-secrets\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.548426 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsll5\" (UniqueName: \"kubernetes.io/projected/9f88d0e7-a6b2-469d-98c6-86747dc3c112-kube-api-access-dsll5\") pod \"console-operator-58897d9998-49tww\" (UID: \"9f88d0e7-a6b2-469d-98c6-86747dc3c112\") " pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.566909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/784c7366-4335-438f-9cc7-6ed4580fc0a5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-bq885\" (UID: \"784c7366-4335-438f-9cc7-6ed4580fc0a5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.577149 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x"] Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.583457 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfhtm\" (UniqueName: \"kubernetes.io/projected/dd1a5521-45f9-4377-970e-4710f6749fa5-kube-api-access-lfhtm\") pod \"multus-admission-controller-857f4d67dd-zbzr7\" (UID: \"dd1a5521-45f9-4377-970e-4710f6749fa5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: W1126 06:27:38.601581 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod283bf004_b2c0_4c8c_b979_a884cf514ab2.slice/crio-85e5f9a154a3d4fb67e814f22b002e8dad663605371dbdb856e800e0c2a5d58a WatchSource:0}: Error finding container 85e5f9a154a3d4fb67e814f22b002e8dad663605371dbdb856e800e0c2a5d58a: Status 404 returned error can't find the container with id 85e5f9a154a3d4fb67e814f22b002e8dad663605371dbdb856e800e0c2a5d58a Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.602974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603154 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2a863afc-708f-4afe-b3f4-c6cb950c23a2-certs\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603178 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6c78\" (UniqueName: \"kubernetes.io/projected/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-kube-api-access-d6c78\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603200 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-ca\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-registration-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603243 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-signing-key\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603264 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-service-ca\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603280 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82rkd\" (UniqueName: \"kubernetes.io/projected/764b5255-c5f2-4e67-8ad9-0050044df158-kube-api-access-82rkd\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x78j7\" (UniqueName: \"kubernetes.io/projected/eceaefdb-3bc4-4199-815f-c5cca21b8110-kube-api-access-x78j7\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1c14a59-51f9-475d-8867-aefb952b66dc-metrics-tls\") pod \"dns-operator-744455d44c-gt6r8\" (UID: \"f1c14a59-51f9-475d-8867-aefb952b66dc\") " pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603334 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c83b476-9e46-440e-be8b-76e02c398f70-serving-cert\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-plugins-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603365 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-config\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603382 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eceaefdb-3bc4-4199-815f-c5cca21b8110-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-socket-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjvnc\" (UniqueName: \"kubernetes.io/projected/2a863afc-708f-4afe-b3f4-c6cb950c23a2-kube-api-access-jjvnc\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-csi-data-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603454 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a969da6-616f-4f99-8271-5a0d6a39b06c-metrics-tls\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603469 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-mountpoint-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603483 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c83b476-9e46-440e-be8b-76e02c398f70-config\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603497 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfb25e56-d6ab-4e49-86de-9a20ba2ee02a-cert\") pod \"ingress-canary-rp4jq\" (UID: \"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a\") " pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603529 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szxmw\" (UniqueName: \"kubernetes.io/projected/8a969da6-616f-4f99-8271-5a0d6a39b06c-kube-api-access-szxmw\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603553 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-tmpfs\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603567 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2a863afc-708f-4afe-b3f4-c6cb950c23a2-node-bootstrap-token\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-apiservice-cert\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603613 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl5n7\" (UniqueName: \"kubernetes.io/projected/f1365993-a452-4e00-92e0-64e1c08ce1c2-kube-api-access-bl5n7\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b0fb0ce-6df9-4e4c-8d48-b35699664601-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vrt6m\" (UID: \"6b0fb0ce-6df9-4e4c-8d48-b35699664601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603658 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46jg9\" (UniqueName: \"kubernetes.io/projected/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-kube-api-access-46jg9\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603672 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-955ln\" (UniqueName: \"kubernetes.io/projected/6b0fb0ce-6df9-4e4c-8d48-b35699664601-kube-api-access-955ln\") pod \"package-server-manager-789f6589d5-vrt6m\" (UID: \"6b0fb0ce-6df9-4e4c-8d48-b35699664601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603694 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc93554a-637c-4af3-a239-bdfab76f780f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603784 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc93554a-637c-4af3-a239-bdfab76f780f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603802 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/de5015a4-08fc-447e-86da-8341572f7143-srv-cert\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603822 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhgx9\" (UniqueName: \"kubernetes.io/projected/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-kube-api-access-dhgx9\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603838 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xglbk\" (UniqueName: \"kubernetes.io/projected/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-kube-api-access-xglbk\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-948ls\" (UniqueName: \"kubernetes.io/projected/f1c14a59-51f9-475d-8867-aefb952b66dc-kube-api-access-948ls\") pod \"dns-operator-744455d44c-gt6r8\" (UID: \"f1c14a59-51f9-475d-8867-aefb952b66dc\") " pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603867 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a969da6-616f-4f99-8271-5a0d6a39b06c-trusted-ca\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-config-volume\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603898 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkxhc\" (UniqueName: \"kubernetes.io/projected/cfb25e56-d6ab-4e49-86de-9a20ba2ee02a-kube-api-access-tkxhc\") pod \"ingress-canary-rp4jq\" (UID: \"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a\") " pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603921 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc93554a-637c-4af3-a239-bdfab76f780f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603959 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kfs2\" (UniqueName: \"kubernetes.io/projected/54b5883f-4f48-4b9e-84b7-46eeb452aedc-kube-api-access-6kfs2\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a11b6fbe-aabd-45e3-9302-3b3678bba644-metrics-tls\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.603992 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-webhook-cert\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wknk\" (UniqueName: \"kubernetes.io/projected/a11b6fbe-aabd-45e3-9302-3b3678bba644-kube-api-access-2wknk\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604029 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eceaefdb-3bc4-4199-815f-c5cca21b8110-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604047 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-srv-cert\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604062 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-signing-cabundle\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntf5k\" (UniqueName: \"kubernetes.io/projected/de5015a4-08fc-447e-86da-8341572f7143-kube-api-access-ntf5k\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.604093 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.104073689 +0000 UTC m=+142.465377641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604134 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a11b6fbe-aabd-45e3-9302-3b3678bba644-config-volume\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-client\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604206 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wns7g\" (UniqueName: \"kubernetes.io/projected/3c83b476-9e46-440e-be8b-76e02c398f70-kube-api-access-wns7g\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604226 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54b5883f-4f48-4b9e-84b7-46eeb452aedc-serving-cert\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-secret-volume\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604277 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a969da6-616f-4f99-8271-5a0d6a39b06c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604313 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/de5015a4-08fc-447e-86da-8341572f7143-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.606674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8a969da6-616f-4f99-8271-5a0d6a39b06c-trusted-ca\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.607351 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-registration-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.607571 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-csi-data-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.607857 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-ca\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.607987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/de5015a4-08fc-447e-86da-8341572f7143-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.608299 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2a863afc-708f-4afe-b3f4-c6cb950c23a2-certs\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.609011 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82hfb\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-kube-api-access-82hfb\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.609616 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-apiservice-cert\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.610679 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a11b6fbe-aabd-45e3-9302-3b3678bba644-metrics-tls\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.613294 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eceaefdb-3bc4-4199-815f-c5cca21b8110-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.613489 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-signing-key\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.614005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-srv-cert\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.614599 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-mountpoint-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.614891 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-secret-volume\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.615119 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c83b476-9e46-440e-be8b-76e02c398f70-config\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.615265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a969da6-616f-4f99-8271-5a0d6a39b06c-metrics-tls\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.615602 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-webhook-cert\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.616244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2a863afc-708f-4afe-b3f4-c6cb950c23a2-node-bootstrap-token\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.616358 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-signing-cabundle\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.616734 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc93554a-637c-4af3-a239-bdfab76f780f-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.617380 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f1c14a59-51f9-475d-8867-aefb952b66dc-metrics-tls\") pod \"dns-operator-744455d44c-gt6r8\" (UID: \"f1c14a59-51f9-475d-8867-aefb952b66dc\") " pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.617553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc93554a-637c-4af3-a239-bdfab76f780f-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.618302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a11b6fbe-aabd-45e3-9302-3b3678bba644-config-volume\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.604164 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-tmpfs\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.619577 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b0fb0ce-6df9-4e4c-8d48-b35699664601-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-vrt6m\" (UID: \"6b0fb0ce-6df9-4e4c-8d48-b35699664601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.619686 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.619922 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-config-volume\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.620058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54b5883f-4f48-4b9e-84b7-46eeb452aedc-serving-cert\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.620244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-socket-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.620286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/764b5255-c5f2-4e67-8ad9-0050044df158-plugins-dir\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.620372 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-config\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.621205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eceaefdb-3bc4-4199-815f-c5cca21b8110-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.621494 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.621649 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfb25e56-d6ab-4e49-86de-9a20ba2ee02a-cert\") pod \"ingress-canary-rp4jq\" (UID: \"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a\") " pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.622317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.624351 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/de5015a4-08fc-447e-86da-8341572f7143-srv-cert\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.608387 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-service-ca\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.626358 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54b5883f-4f48-4b9e-84b7-46eeb452aedc-etcd-client\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.626458 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c83b476-9e46-440e-be8b-76e02c398f70-serving-cert\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.626474 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.633620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpfb8\" (UniqueName: \"kubernetes.io/projected/cb8fd937-d759-4e21-98af-5b833a8b3f52-kube-api-access-wpfb8\") pod \"control-plane-machine-set-operator-78cbb6b69f-x5x92\" (UID: \"cb8fd937-d759-4e21-98af-5b833a8b3f52\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.641670 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcxd4\" (UniqueName: \"kubernetes.io/projected/3ae16d02-bfb3-4776-b438-ce9e8ed409ec-kube-api-access-rcxd4\") pod \"openshift-controller-manager-operator-756b6f6bc6-2lfgm\" (UID: \"3ae16d02-bfb3-4776-b438-ce9e8ed409ec\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.658671 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.661987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l578j\" (UniqueName: \"kubernetes.io/projected/3316a24f-1cce-4358-a8b4-03b40f9ceec0-kube-api-access-l578j\") pod \"migrator-59844c95c7-rb4zf\" (UID: \"3316a24f-1cce-4358-a8b4-03b40f9ceec0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.665631 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.685074 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-bound-sa-token\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.693371 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.702448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cskh\" (UniqueName: \"kubernetes.io/projected/e1cb46aa-dd22-49d0-9a14-dab875c02723-kube-api-access-4cskh\") pod \"machine-config-controller-84d6567774-q5vmb\" (UID: \"e1cb46aa-dd22-49d0-9a14-dab875c02723\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.706009 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.706360 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.206343201 +0000 UTC m=+142.567647153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.722646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e9851fb-25bb-4dbe-a387-3c187bc4faeb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nmzpp\" (UID: \"9e9851fb-25bb-4dbe-a387-3c187bc4faeb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.731534 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.740080 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.746376 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8zrj\" (UniqueName: \"kubernetes.io/projected/4d70bf4c-6643-45d9-8b43-aab969fe7bfa-kube-api-access-k8zrj\") pod \"openshift-apiserver-operator-796bbdcf4f-d9bcm\" (UID: \"4d70bf4c-6643-45d9-8b43-aab969fe7bfa\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.747726 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.751079 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj"] Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.803314 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82rkd\" (UniqueName: \"kubernetes.io/projected/764b5255-c5f2-4e67-8ad9-0050044df158-kube-api-access-82rkd\") pod \"csi-hostpathplugin-pv8cb\" (UID: \"764b5255-c5f2-4e67-8ad9-0050044df158\") " pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.808475 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.808957 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.308942723 +0000 UTC m=+142.670246675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.830013 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntf5k\" (UniqueName: \"kubernetes.io/projected/de5015a4-08fc-447e-86da-8341572f7143-kube-api-access-ntf5k\") pod \"olm-operator-6b444d44fb-bntf8\" (UID: \"de5015a4-08fc-447e-86da-8341572f7143\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.832688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6c78\" (UniqueName: \"kubernetes.io/projected/05abb732-1cb3-4e8b-817e-7bdc53d29c7f-kube-api-access-d6c78\") pod \"packageserver-d55dfcdfc-kz9zl\" (UID: \"05abb732-1cb3-4e8b-817e-7bdc53d29c7f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.853398 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wns7g\" (UniqueName: \"kubernetes.io/projected/3c83b476-9e46-440e-be8b-76e02c398f70-kube-api-access-wns7g\") pod \"service-ca-operator-777779d784-szmvm\" (UID: \"3c83b476-9e46-440e-be8b-76e02c398f70\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.868909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjvnc\" (UniqueName: \"kubernetes.io/projected/2a863afc-708f-4afe-b3f4-c6cb950c23a2-kube-api-access-jjvnc\") pod \"machine-config-server-lt2r2\" (UID: \"2a863afc-708f-4afe-b3f4-c6cb950c23a2\") " pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.889883 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x78j7\" (UniqueName: \"kubernetes.io/projected/eceaefdb-3bc4-4199-815f-c5cca21b8110-kube-api-access-x78j7\") pod \"kube-storage-version-migrator-operator-b67b599dd-hz54h\" (UID: \"eceaefdb-3bc4-4199-815f-c5cca21b8110\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.897337 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cv6pw"] Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.897674 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.910981 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:38 crc kubenswrapper[4775]: E1126 06:27:38.911420 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.411410021 +0000 UTC m=+142.772713973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.913015 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.917969 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl5n7\" (UniqueName: \"kubernetes.io/projected/f1365993-a452-4e00-92e0-64e1c08ce1c2-kube-api-access-bl5n7\") pod \"marketplace-operator-79b997595-z59ck\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.931102 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.937695 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lt2r2" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.948360 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kfs2\" (UniqueName: \"kubernetes.io/projected/54b5883f-4f48-4b9e-84b7-46eeb452aedc-kube-api-access-6kfs2\") pod \"etcd-operator-b45778765-rtsmv\" (UID: \"54b5883f-4f48-4b9e-84b7-46eeb452aedc\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.950243 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wknk\" (UniqueName: \"kubernetes.io/projected/a11b6fbe-aabd-45e3-9302-3b3678bba644-kube-api-access-2wknk\") pod \"dns-default-9cq9z\" (UID: \"a11b6fbe-aabd-45e3-9302-3b3678bba644\") " pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.951803 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.976665 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cc93554a-637c-4af3-a239-bdfab76f780f-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9czjc\" (UID: \"cc93554a-637c-4af3-a239-bdfab76f780f\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.982045 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8"] Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.985138 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46jg9\" (UniqueName: \"kubernetes.io/projected/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-kube-api-access-46jg9\") pod \"collect-profiles-29402295-bqjbg\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:38 crc kubenswrapper[4775]: I1126 06:27:38.991023 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4bw62"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.007371 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-955ln\" (UniqueName: \"kubernetes.io/projected/6b0fb0ce-6df9-4e4c-8d48-b35699664601-kube-api-access-955ln\") pod \"package-server-manager-789f6589d5-vrt6m\" (UID: \"6b0fb0ce-6df9-4e4c-8d48-b35699664601\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.019495 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.019828 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.519811813 +0000 UTC m=+142.881115765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.062023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.062683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a969da6-616f-4f99-8271-5a0d6a39b06c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.064451 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.068723 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.068741 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkxhc\" (UniqueName: \"kubernetes.io/projected/cfb25e56-d6ab-4e49-86de-9a20ba2ee02a-kube-api-access-tkxhc\") pod \"ingress-canary-rp4jq\" (UID: \"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a\") " pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.068793 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.074810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szxmw\" (UniqueName: \"kubernetes.io/projected/8a969da6-616f-4f99-8271-5a0d6a39b06c-kube-api-access-szxmw\") pod \"ingress-operator-5b745b69d9-t29hf\" (UID: \"8a969da6-616f-4f99-8271-5a0d6a39b06c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.077497 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.086403 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.109262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.110944 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.116793 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.118100 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xglbk\" (UniqueName: \"kubernetes.io/projected/6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab-kube-api-access-xglbk\") pod \"catalog-operator-68c6474976-db5pc\" (UID: \"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.122670 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.123018 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.623007671 +0000 UTC m=+142.984311623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.129692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.132194 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhgx9\" (UniqueName: \"kubernetes.io/projected/e90134d4-f1be-49d0-a01d-b5d2f7b3c151-kube-api-access-dhgx9\") pod \"service-ca-9c57cc56f-n4dwr\" (UID: \"e90134d4-f1be-49d0-a01d-b5d2f7b3c151\") " pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.137242 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-948ls\" (UniqueName: \"kubernetes.io/projected/f1c14a59-51f9-475d-8867-aefb952b66dc-kube-api-access-948ls\") pod \"dns-operator-744455d44c-gt6r8\" (UID: \"f1c14a59-51f9-475d-8867-aefb952b66dc\") " pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.139758 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.147002 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.159755 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.181459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.191956 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-49tww"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.209979 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rp4jq" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.224093 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.226170 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.226682 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.726665082 +0000 UTC m=+143.087969034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.231597 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf"] Nov 26 06:27:39 crc kubenswrapper[4775]: W1126 06:27:39.267910 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb8fd937_d759_4e21_98af_5b833a8b3f52.slice/crio-9e179b7a58bfb1a3bcb776154ab47c8646ec69bc605c0722a11d03e420cfa9e7 WatchSource:0}: Error finding container 9e179b7a58bfb1a3bcb776154ab47c8646ec69bc605c0722a11d03e420cfa9e7: Status 404 returned error can't find the container with id 9e179b7a58bfb1a3bcb776154ab47c8646ec69bc605c0722a11d03e420cfa9e7 Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.289063 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" event={"ID":"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b","Type":"ContainerStarted","Data":"7ea6538a8264070a5f6626fed13789e517adff3a0cd6d2c662fc0797c2c968e1"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.291631 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pv8cb"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.322061 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.324983 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" event={"ID":"ed720241-da26-4f88-9838-c851d15fd2d8","Type":"ContainerStarted","Data":"b03ae9da60b46c4780d68fa3b535700b2a8483c5597d1abf17cd8bcfe8f5752b"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.331597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.331870 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.831859216 +0000 UTC m=+143.193163168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.348573 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" event={"ID":"283bf004-b2c0-4c8c-b979-a884cf514ab2","Type":"ContainerStarted","Data":"22d5431b8ac178e31d4b056e2ffa22a2bd0f36d86c70610d8d1c8c6dd80f9b60"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.348622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" event={"ID":"283bf004-b2c0-4c8c-b979-a884cf514ab2","Type":"ContainerStarted","Data":"0150935d28120510b7fb0c3cc4d1d68ec9c87f8231b3ab1a757f9a652baf110b"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.348631 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" event={"ID":"283bf004-b2c0-4c8c-b979-a884cf514ab2","Type":"ContainerStarted","Data":"85e5f9a154a3d4fb67e814f22b002e8dad663605371dbdb856e800e0c2a5d58a"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.349991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" event={"ID":"74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb","Type":"ContainerStarted","Data":"90ee929c34caf1b792ea1c2b12a3c21e5c433bfac0e9a24374a4e00c765eae6c"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.356103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" event={"ID":"9a2a9335-a9d7-4145-af84-2084ad5d8c14","Type":"ContainerStarted","Data":"e865abde9bc17fc04d6a03164d3c015111c20dd1ae289b8f2e7cc9dccc466977"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.365076 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-zbzr7"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.379540 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" event={"ID":"61d701be-8e8d-4d60-a562-328010fc2003","Type":"ContainerStarted","Data":"a5e22261a7db7e32d5607a55a8fd75f95373c95c0b2d671f90456e415682bc20"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.383705 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.393224 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-drckl" event={"ID":"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54","Type":"ContainerStarted","Data":"04acf473ec453efaa3c1c1e08735fcd315a3548746d1b8fb68e23f2e2e7d3cfb"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.393258 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-drckl" event={"ID":"f02e8337-4b5c-4ec3-b0a2-38d674cbfe54","Type":"ContainerStarted","Data":"25dcb84a9ab5554d7230ed678c393c81ad42796386c30f1e96363cbf396b8265"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.399117 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.408366 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.410054 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.410495 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" event={"ID":"23e94d9b-5940-42b3-9876-17fc7bf7fa18","Type":"ContainerStarted","Data":"dce60818dfc6f820604f341f8157ffb42583934a1961dd976032f921c273c7e1"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.410516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" event={"ID":"23e94d9b-5940-42b3-9876-17fc7bf7fa18","Type":"ContainerStarted","Data":"b6ec45319cdcf2bf7fb8a38e2c1331788944a8892620594f07b744344ba0f926"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.410623 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:39 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:39 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:39 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.410655 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.412045 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lt2r2" event={"ID":"2a863afc-708f-4afe-b3f4-c6cb950c23a2","Type":"ContainerStarted","Data":"d3600915db253e5e4cf0497a4d2def2fa77d5456792d31b92f0276e0693a3fa5"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.423858 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vl6j8" event={"ID":"268d85ad-f165-4cd2-931f-1da2caefe3cd","Type":"ContainerStarted","Data":"8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.423909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vl6j8" event={"ID":"268d85ad-f165-4cd2-931f-1da2caefe3cd","Type":"ContainerStarted","Data":"c6e390cd4c627d8b833dd438f50c7494f534145c2887be6dbcd9b0f176c89724"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.433773 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.434293 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.934277623 +0000 UTC m=+143.295581575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.435643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.437631 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:39.937622265 +0000 UTC m=+143.298926217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.469263 4775 generic.go:334] "Generic (PLEG): container finished" podID="e0344f36-af1e-44ad-9f72-7a616d6e222c" containerID="54c8ee3565ea523dcd02e67aae3e66726a91288916e3ceaf95e183ba003334b6" exitCode=0 Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.469338 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kclct" event={"ID":"e0344f36-af1e-44ad-9f72-7a616d6e222c","Type":"ContainerDied","Data":"54c8ee3565ea523dcd02e67aae3e66726a91288916e3ceaf95e183ba003334b6"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.487087 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" event={"ID":"ab5f6914-f92e-47d6-9810-b61ce5151538","Type":"ContainerStarted","Data":"5cf062586d6788f80095c119f8d8701c8e548e7944b9686b78677e97d455693c"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.487857 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.493849 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.496294 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" event={"ID":"4942219a-29bb-4f6b-904e-e54d0222d948","Type":"ContainerStarted","Data":"abe4d6ea1c063d46207c4c6b0f6ab1b24c5c3e4ffbaf99a1ada933974b802bf4"} Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.498816 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-44mw9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.498883 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-44mw9" podUID="2278df09-65b1-440e-a613-7bea0b8c73cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.503534 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.536790 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.539123 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.039105026 +0000 UTC m=+143.400408978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.638296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.638647 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.138635543 +0000 UTC m=+143.499939495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.683378 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.695845 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.738559 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.738960 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.739779 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.239750154 +0000 UTC m=+143.601054106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.752857 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-szmvm"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.788739 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.812203 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rtsmv"] Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.843026 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.843388 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.343373424 +0000 UTC m=+143.704677376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.902882 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-44mw9" podStartSLOduration=122.902868236 podStartE2EDuration="2m2.902868236s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:39.901814787 +0000 UTC m=+143.263118739" watchObservedRunningTime="2025-11-26 06:27:39.902868236 +0000 UTC m=+143.264172188" Nov 26 06:27:39 crc kubenswrapper[4775]: I1126 06:27:39.949289 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:39 crc kubenswrapper[4775]: E1126 06:27:39.949616 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.449600836 +0000 UTC m=+143.810904788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.032448 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-p4qsm" podStartSLOduration=122.032254017 podStartE2EDuration="2m2.032254017s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.028235346 +0000 UTC m=+143.389539308" watchObservedRunningTime="2025-11-26 06:27:40.032254017 +0000 UTC m=+143.393557979" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.050457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.051138 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.551124288 +0000 UTC m=+143.912428240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.158738 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.159169 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.65915358 +0000 UTC m=+144.020457532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.261423 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.261964 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.761952637 +0000 UTC m=+144.123256579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.327216 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vl6j8" podStartSLOduration=123.327202958 podStartE2EDuration="2m3.327202958s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.325983784 +0000 UTC m=+143.687287736" watchObservedRunningTime="2025-11-26 06:27:40.327202958 +0000 UTC m=+143.688506910" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.365336 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.371505 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.372065 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:40.872051946 +0000 UTC m=+144.233355898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.418689 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.439576 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:40 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:40 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:40 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.439642 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.464379 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" podStartSLOduration=122.464347062 podStartE2EDuration="2m2.464347062s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.455108087 +0000 UTC m=+143.816412059" watchObservedRunningTime="2025-11-26 06:27:40.464347062 +0000 UTC m=+143.825651014" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.490641 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.496162 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fxkqj" podStartSLOduration=122.49614796 podStartE2EDuration="2m2.49614796s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.493357163 +0000 UTC m=+143.854661115" watchObservedRunningTime="2025-11-26 06:27:40.49614796 +0000 UTC m=+143.857451912" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.509401 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.517081 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.017036956 +0000 UTC m=+144.378340908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.534278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" event={"ID":"764b5255-c5f2-4e67-8ad9-0050044df158","Type":"ContainerStarted","Data":"366a268455f5c0558461bdf7e0c3c1fb476c48897249e27ecc13ef31969340e6"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.541108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" event={"ID":"3ae16d02-bfb3-4776-b438-ce9e8ed409ec","Type":"ContainerStarted","Data":"7507e6a5ebede55d47ac6b69036c0a2a3d7e317cbf63986c64c40dbed140e5dc"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.541893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" event={"ID":"784c7366-4335-438f-9cc7-6ed4580fc0a5","Type":"ContainerStarted","Data":"f48485fb75feb8a77f8b1d0475639b8a47f318023518961636c4422f13a790d0"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.549889 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-9cq9z"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.570997 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" event={"ID":"de5015a4-08fc-447e-86da-8341572f7143","Type":"ContainerStarted","Data":"632e76415620956f42e5070043b2ee0f086d06135be543fcc18a82d261742cbe"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.587955 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-49tww" event={"ID":"9f88d0e7-a6b2-469d-98c6-86747dc3c112","Type":"ContainerStarted","Data":"d7f2588900b14ca0ed5e38a98b055ed73477af46cdddc511b611e876ac5d57ff"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.600433 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.602876 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" podStartSLOduration=123.602859165 podStartE2EDuration="2m3.602859165s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.593309632 +0000 UTC m=+143.954613614" watchObservedRunningTime="2025-11-26 06:27:40.602859165 +0000 UTC m=+143.964163117" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.616694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.617242 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.117228372 +0000 UTC m=+144.478532324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.619363 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n4dwr"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.627081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" event={"ID":"e1cb46aa-dd22-49d0-9a14-dab875c02723","Type":"ContainerStarted","Data":"651cb82976dff27077cebcff408101657d35d7b2892232fd25db9d1721beac1b"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.627118 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" event={"ID":"e1cb46aa-dd22-49d0-9a14-dab875c02723","Type":"ContainerStarted","Data":"6c8c76ef47b317379a95bcbc70f9e3595b03ba7a67b82574836bea15687bac84"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.639229 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" event={"ID":"54b5883f-4f48-4b9e-84b7-46eeb452aedc","Type":"ContainerStarted","Data":"9a16c8f332832960701bb6cfabe0ebc900fe520a230e3ae4ff5e8fed0330610b"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.690149 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" event={"ID":"9e9851fb-25bb-4dbe-a387-3c187bc4faeb","Type":"ContainerStarted","Data":"852b9a90cfab2ae40c17077eb7aa09e910e9ca74da09386ea1ff351eb06c674f"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.718054 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-drckl" podStartSLOduration=122.718036964 podStartE2EDuration="2m2.718036964s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.717501059 +0000 UTC m=+144.078805011" watchObservedRunningTime="2025-11-26 06:27:40.718036964 +0000 UTC m=+144.079340926" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.724408 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.724845 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.224835392 +0000 UTC m=+144.586139344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: W1126 06:27:40.799176 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d1c89a1_5fc7_46f7_b1b6_b2171abbe3ab.slice/crio-10351bd91e3d80ecb98354a411e595f657563627bf47bda9646c7ab669dc96c6 WatchSource:0}: Error finding container 10351bd91e3d80ecb98354a411e595f657563627bf47bda9646c7ab669dc96c6: Status 404 returned error can't find the container with id 10351bd91e3d80ecb98354a411e595f657563627bf47bda9646c7ab669dc96c6 Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.800500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" event={"ID":"b188fc55-2f3e-40a2-a779-863cacc2a1eb","Type":"ContainerStarted","Data":"5f8ff4925f70b2c11737a119ea49db5d244f0793ebee4b8ae327dddc83da71cc"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.800549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" event={"ID":"b188fc55-2f3e-40a2-a779-863cacc2a1eb","Type":"ContainerStarted","Data":"26d2309d58879f43158d3a8d55e608a5ad115b86f34fbf6313a182f433f23277"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.834785 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.836052 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.33601344 +0000 UTC m=+144.697317382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.836218 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.836511 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.336498404 +0000 UTC m=+144.697802356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.840339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" event={"ID":"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3","Type":"ContainerStarted","Data":"d3296732f9660e24bedf1d12c43b482380835856db689327eca81c33c2e3a16c"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.849969 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" event={"ID":"9a2a9335-a9d7-4145-af84-2084ad5d8c14","Type":"ContainerStarted","Data":"f7aae7c34718251fa80fff72c1d16136f4b870feb321956d5b3d849075d3a048"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.876389 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" podStartSLOduration=122.876373864 podStartE2EDuration="2m2.876373864s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.867263723 +0000 UTC m=+144.228567675" watchObservedRunningTime="2025-11-26 06:27:40.876373864 +0000 UTC m=+144.237677826" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.895463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" event={"ID":"dd1a5521-45f9-4377-970e-4710f6749fa5","Type":"ContainerStarted","Data":"c7c2eaee0e5c902954fcb5789833cb10dcd535565f0f7fb5a084efb00a9dd07f"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.903932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" event={"ID":"61d701be-8e8d-4d60-a562-328010fc2003","Type":"ContainerStarted","Data":"31822ae474dcc4eee059f2f0422b6306e243f86f5f35c6c6113d280ff03c89d1"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.904642 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.936853 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.937376 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.937410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" event={"ID":"b4a0f0cb-cf20-4d5a-b261-4aedd3459e2b","Type":"ContainerStarted","Data":"11e71b88ffc77bd398118f3ac24bf396e5d0b1e173cdda6217675499b0c566ad"} Nov 26 06:27:40 crc kubenswrapper[4775]: E1126 06:27:40.937611 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.437598564 +0000 UTC m=+144.798902516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.942497 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-26sfm" podStartSLOduration=124.942477099 podStartE2EDuration="2m4.942477099s" podCreationTimestamp="2025-11-26 06:25:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.92912483 +0000 UTC m=+144.290428782" watchObservedRunningTime="2025-11-26 06:27:40.942477099 +0000 UTC m=+144.303781051" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.956040 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" event={"ID":"3c83b476-9e46-440e-be8b-76e02c398f70","Type":"ContainerStarted","Data":"9d5bd518183add3d75ec345ba314d3c0841ddadeef5bdb472a18f82ffb1dc1d7"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.959523 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z59ck"] Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.962121 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vs44x" podStartSLOduration=122.9621074 podStartE2EDuration="2m2.9621074s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:40.961124643 +0000 UTC m=+144.322428595" watchObservedRunningTime="2025-11-26 06:27:40.9621074 +0000 UTC m=+144.323411352" Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.964390 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" event={"ID":"cb8fd937-d759-4e21-98af-5b833a8b3f52","Type":"ContainerStarted","Data":"9e179b7a58bfb1a3bcb776154ab47c8646ec69bc605c0722a11d03e420cfa9e7"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.973048 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" event={"ID":"3316a24f-1cce-4358-a8b4-03b40f9ceec0","Type":"ContainerStarted","Data":"eaf930f37a09adca41abfc282fcb3d7f70a8ed77339b4a51c7b0eb77f715580c"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.998150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" event={"ID":"4d70bf4c-6643-45d9-8b43-aab969fe7bfa","Type":"ContainerStarted","Data":"3d48e853fe543f926828974748996a7ecb91f1a9b3c483f9ad7021f2b3d44da8"} Nov 26 06:27:40 crc kubenswrapper[4775]: I1126 06:27:40.998188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" event={"ID":"4d70bf4c-6643-45d9-8b43-aab969fe7bfa","Type":"ContainerStarted","Data":"b6a87e0ddb6cca2d80924383de68594a5abb78be06d8b161b15cfd2aec0d034f"} Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:40.998405 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-44mw9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:40.998451 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-44mw9" podUID="2278df09-65b1-440e-a613-7bea0b8c73cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.006934 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf"] Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.055632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.059938 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.55992392 +0000 UTC m=+144.921227872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.085933 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rp4jq"] Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.092238 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gt6r8"] Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.120235 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-d9bcm" podStartSLOduration=124.120219155 podStartE2EDuration="2m4.120219155s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:41.118836556 +0000 UTC m=+144.480140508" watchObservedRunningTime="2025-11-26 06:27:41.120219155 +0000 UTC m=+144.481523107" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.169686 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.169978 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.669938287 +0000 UTC m=+145.031242239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.176986 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" podStartSLOduration=123.176968681 podStartE2EDuration="2m3.176968681s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:41.143033644 +0000 UTC m=+144.504337596" watchObservedRunningTime="2025-11-26 06:27:41.176968681 +0000 UTC m=+144.538272633" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.177452 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" podStartSLOduration=123.177447744 podStartE2EDuration="2m3.177447744s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:41.169532746 +0000 UTC m=+144.530836698" watchObservedRunningTime="2025-11-26 06:27:41.177447744 +0000 UTC m=+144.538751716" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.221225 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" podStartSLOduration=124.221207062 podStartE2EDuration="2m4.221207062s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:41.1957975 +0000 UTC m=+144.557101452" watchObservedRunningTime="2025-11-26 06:27:41.221207062 +0000 UTC m=+144.582511024" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.222406 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lt2r2" podStartSLOduration=5.222399955 podStartE2EDuration="5.222399955s" podCreationTimestamp="2025-11-26 06:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:41.2204207 +0000 UTC m=+144.581724652" watchObservedRunningTime="2025-11-26 06:27:41.222399955 +0000 UTC m=+144.583703907" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.266747 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4bw62" podStartSLOduration=124.266697957 podStartE2EDuration="2m4.266697957s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:41.266287196 +0000 UTC m=+144.627591148" watchObservedRunningTime="2025-11-26 06:27:41.266697957 +0000 UTC m=+144.628001909" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.277200 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.277494 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.777483125 +0000 UTC m=+145.138787077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.328928 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.379174 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.379466 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.879451959 +0000 UTC m=+145.240755911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.417579 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:41 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:41 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:41 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.417809 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.480460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.481012 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:41.981001152 +0000 UTC m=+145.342305094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.583328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.584256 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.084240412 +0000 UTC m=+145.445544364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.716546 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.717029 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.217011936 +0000 UTC m=+145.578315888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.822491 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.823144 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.323128435 +0000 UTC m=+145.684432377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:41 crc kubenswrapper[4775]: I1126 06:27:41.924685 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:41 crc kubenswrapper[4775]: E1126 06:27:41.925311 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.425296375 +0000 UTC m=+145.786600327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.026285 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.026789 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.526767516 +0000 UTC m=+145.888071468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.050074 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" event={"ID":"784c7366-4335-438f-9cc7-6ed4580fc0a5","Type":"ContainerStarted","Data":"0553b311795eae3e5d2fcbb36f822672f620d170b1ff8ab8cf8d9da9a50ee96c"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.075421 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-bq885" podStartSLOduration=124.075407298 podStartE2EDuration="2m4.075407298s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.074901574 +0000 UTC m=+145.436205526" watchObservedRunningTime="2025-11-26 06:27:42.075407298 +0000 UTC m=+145.436711250" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.082300 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-x5x92" event={"ID":"cb8fd937-d759-4e21-98af-5b833a8b3f52","Type":"ContainerStarted","Data":"fa153288622587d8d4747b66c96fb27474f4cdc5da19554019230be8b9981647"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.099172 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" event={"ID":"8a969da6-616f-4f99-8271-5a0d6a39b06c","Type":"ContainerStarted","Data":"c6fbdee1fc85a19487bf3b0c24ba2ac983c269007ce2c87550f818940074386f"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.099215 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" event={"ID":"8a969da6-616f-4f99-8271-5a0d6a39b06c","Type":"ContainerStarted","Data":"614419cc14e4dda4f7f70da94291415b14b225e81b03e6069faac5c31e813f7c"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.107143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" event={"ID":"05abb732-1cb3-4e8b-817e-7bdc53d29c7f","Type":"ContainerStarted","Data":"133c8748dc859efc5128d86a37b4a803c3cf7d70bd8efff8d142453a7b0ba0a6"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.107186 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" event={"ID":"05abb732-1cb3-4e8b-817e-7bdc53d29c7f","Type":"ContainerStarted","Data":"56923634cd842646f6abfe119db442ca1124779e7c60a6483ae56cbecd4f7bff"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.107705 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.110344 4775 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kz9zl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.110386 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" podUID="05abb732-1cb3-4e8b-817e-7bdc53d29c7f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.133150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" event={"ID":"6b0fb0ce-6df9-4e4c-8d48-b35699664601","Type":"ContainerStarted","Data":"56361a39d802fde753f6b3f790ed9e5f76bd85c255607db73b35567b0e9f93ec"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.133193 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" event={"ID":"6b0fb0ce-6df9-4e4c-8d48-b35699664601","Type":"ContainerStarted","Data":"1ed78aec852130d3cebf93bf3e7ba5202dbb41bc0039bbe524d00607a56cecae"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.134328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.135141 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.635124846 +0000 UTC m=+145.996428798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.151886 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" podStartSLOduration=124.151871049 podStartE2EDuration="2m4.151871049s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.150522091 +0000 UTC m=+145.511826033" watchObservedRunningTime="2025-11-26 06:27:42.151871049 +0000 UTC m=+145.513175001" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.169932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-szmvm" event={"ID":"3c83b476-9e46-440e-be8b-76e02c398f70","Type":"ContainerStarted","Data":"b472279132feae5332dd6cd4c39af96f0d9f23f1021a1dd89fe0c7eb26b51c6b"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.208167 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" event={"ID":"f1c14a59-51f9-475d-8867-aefb952b66dc","Type":"ContainerStarted","Data":"5e884e5539ca18c090db4eb6b916eebf072adcc0cad10f2cb845b24fd97f6504"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.218078 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.218232 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.233750 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" event={"ID":"764b5255-c5f2-4e67-8ad9-0050044df158","Type":"ContainerStarted","Data":"c04eab142f6d1a20d2dd2f00c54e472fb2ef15633a3d0bc80d9b7851ae623efb"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.235189 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.236089 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.736072673 +0000 UTC m=+146.097376625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.237850 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.307416 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" event={"ID":"e90134d4-f1be-49d0-a01d-b5d2f7b3c151","Type":"ContainerStarted","Data":"db535ab4ed0d90ac4a46eafcfcb11a1c8b44171bcecf07e5bf60979c77225631"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.307506 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" event={"ID":"e90134d4-f1be-49d0-a01d-b5d2f7b3c151","Type":"ContainerStarted","Data":"81bdacae6d2d0b61740ec2a354aad8ebe8ecd29b8d14806b0932516386bd93f4"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.326370 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" event={"ID":"f1365993-a452-4e00-92e0-64e1c08ce1c2","Type":"ContainerStarted","Data":"4c8b9705a97ef316c2823683118950056b723dbb9cf9bf9bad5e7378de964bee"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.327001 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.338745 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z59ck container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.338784 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.339612 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.340665 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.840652659 +0000 UTC m=+146.201956601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.360500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" event={"ID":"de5015a4-08fc-447e-86da-8341572f7143","Type":"ContainerStarted","Data":"acd7a4e2cebb4b515096fa4e1fe4f94180b5e90f4a0b9d3b0816553a9c48b00c"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.363308 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.374423 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" podStartSLOduration=124.374411251 podStartE2EDuration="2m4.374411251s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.373261639 +0000 UTC m=+145.734565591" watchObservedRunningTime="2025-11-26 06:27:42.374411251 +0000 UTC m=+145.735715203" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.375726 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-n4dwr" podStartSLOduration=124.375708567 podStartE2EDuration="2m4.375708567s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.348923367 +0000 UTC m=+145.710227319" watchObservedRunningTime="2025-11-26 06:27:42.375708567 +0000 UTC m=+145.737012509" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.400124 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.408952 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rp4jq" event={"ID":"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a","Type":"ContainerStarted","Data":"9ace86cc01ae463f0dcc951f71c1a58c5a7f0430f9af5d480090cba073aca639"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.419324 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bntf8" podStartSLOduration=124.41930354 podStartE2EDuration="2m4.41930354s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.409640013 +0000 UTC m=+145.770943965" watchObservedRunningTime="2025-11-26 06:27:42.41930354 +0000 UTC m=+145.780607492" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.430026 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:42 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:42 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:42 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.430066 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.432212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" event={"ID":"cc93554a-637c-4af3-a239-bdfab76f780f","Type":"ContainerStarted","Data":"ac159d7381e9836ea2d9ebf449bfd49e4b0fa5a7076cf63dbba18a8dbf5419b2"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.436610 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" event={"ID":"54b5883f-4f48-4b9e-84b7-46eeb452aedc","Type":"ContainerStarted","Data":"1545cb486cdd5484a61007d8fad85b2470f74a47c42939f2098e05d9643ca52b"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.440384 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.441335 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:42.941318197 +0000 UTC m=+146.302622159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.453321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lt2r2" event={"ID":"2a863afc-708f-4afe-b3f4-c6cb950c23a2","Type":"ContainerStarted","Data":"7a88e08a6830dbb665e86dbf0710ac5015772201867d33e0670e19ed42fd1898"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.474265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" event={"ID":"eceaefdb-3bc4-4199-815f-c5cca21b8110","Type":"ContainerStarted","Data":"7b0b903528838edfa1004d34e4ddcb816ec30a68244c1f58ee171c0c4e0ce08e"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.474306 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" event={"ID":"eceaefdb-3bc4-4199-815f-c5cca21b8110","Type":"ContainerStarted","Data":"f1a2abdeead2cb7d0d7709ef7e0f3048423bc59cc11f55286ea12836fd2798ff"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.476010 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" event={"ID":"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab","Type":"ContainerStarted","Data":"22a7207476f7e9181b8f536ed58f1da04f621a75c514648d5f76e96155aabc54"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.476033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" event={"ID":"6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab","Type":"ContainerStarted","Data":"10351bd91e3d80ecb98354a411e595f657563627bf47bda9646c7ab669dc96c6"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.476665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.478196 4775 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-db5pc container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.478232 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" podUID="6d1c89a1-5fc7-46f7-b1b6-b2171abbe3ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.479683 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" event={"ID":"3ae16d02-bfb3-4776-b438-ce9e8ed409ec","Type":"ContainerStarted","Data":"3077c2020864649108b65dd158107d2c0737614fdc72100349f2632dc89c9532"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.480803 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" event={"ID":"9e9851fb-25bb-4dbe-a387-3c187bc4faeb","Type":"ContainerStarted","Data":"f64f8d868fde700d1d677e28e20e6cca10c0590a20b7f05efb5db33c6c2c6baa"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.482055 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" event={"ID":"3316a24f-1cce-4358-a8b4-03b40f9ceec0","Type":"ContainerStarted","Data":"ed5beb6e838f75ab0e87154cd6c7cadccde7bb54f3e72220ed0b53d66df25972"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.482078 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" event={"ID":"3316a24f-1cce-4358-a8b4-03b40f9ceec0","Type":"ContainerStarted","Data":"4a9db737b751f4a410bf89494a797b8099a2fa1d4a6110cc24d7eeeef2082f67"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.515910 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" podStartSLOduration=124.515890336 podStartE2EDuration="2m4.515890336s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.51498073 +0000 UTC m=+145.876284682" watchObservedRunningTime="2025-11-26 06:27:42.515890336 +0000 UTC m=+145.877194288" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.520150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kclct" event={"ID":"e0344f36-af1e-44ad-9f72-7a616d6e222c","Type":"ContainerStarted","Data":"000481f85e2ced59894da81d24a4dfe77d9abc3ba60685cda44d74a865ed99ff"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.544259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.547316 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.047301143 +0000 UTC m=+146.408605095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.552834 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hz54h" podStartSLOduration=124.552820345 podStartE2EDuration="2m4.552820345s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.552242209 +0000 UTC m=+145.913546161" watchObservedRunningTime="2025-11-26 06:27:42.552820345 +0000 UTC m=+145.914124297" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.602469 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rtsmv" podStartSLOduration=124.602452175 podStartE2EDuration="2m4.602452175s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.600496391 +0000 UTC m=+145.961800343" watchObservedRunningTime="2025-11-26 06:27:42.602452175 +0000 UTC m=+145.963756127" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.604003 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" event={"ID":"b188fc55-2f3e-40a2-a779-863cacc2a1eb","Type":"ContainerStarted","Data":"0416654bb3a1a6e4d85fa5c1d5ffa2c44dbd6bce950bbaca8cd2a6380d72794c"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.605162 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.608803 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.608910 4775 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kclct container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.608964 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kclct" podUID="e0344f36-af1e-44ad-9f72-7a616d6e222c" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.623268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9cq9z" event={"ID":"a11b6fbe-aabd-45e3-9302-3b3678bba644","Type":"ContainerStarted","Data":"9626ed2036302ad5a9b37d540503e4e43efa1426b4ca8a5b85850365b73b7956"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.626780 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" event={"ID":"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3","Type":"ContainerStarted","Data":"d95be0bed38af2377ee70712b06f96a07c0956b6629e4603ad5aeee39865f917"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.636138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" event={"ID":"dd1a5521-45f9-4377-970e-4710f6749fa5","Type":"ContainerStarted","Data":"a9af2618686049e535ca00e5e86ce08f668934b60323f460d9626829060ea41b"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.638902 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-49tww" event={"ID":"9f88d0e7-a6b2-469d-98c6-86747dc3c112","Type":"ContainerStarted","Data":"b96a8b021a4577ce936672ccb6c66edececeb551edfac67b144af1a492d15249"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.639528 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.642396 4775 generic.go:334] "Generic (PLEG): container finished" podID="9a2a9335-a9d7-4145-af84-2084ad5d8c14" containerID="f7aae7c34718251fa80fff72c1d16136f4b870feb321956d5b3d849075d3a048" exitCode=0 Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.642517 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" event={"ID":"9a2a9335-a9d7-4145-af84-2084ad5d8c14","Type":"ContainerDied","Data":"f7aae7c34718251fa80fff72c1d16136f4b870feb321956d5b3d849075d3a048"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.642684 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" event={"ID":"9a2a9335-a9d7-4145-af84-2084ad5d8c14","Type":"ContainerStarted","Data":"da899dbb8f967a704eb6eba4425d9b771c48aebf2ca3f560b11171bb4d2dbdb2"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.642843 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.678420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.678616 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kclct" podStartSLOduration=125.678603346 podStartE2EDuration="2m5.678603346s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.678405381 +0000 UTC m=+146.039709343" watchObservedRunningTime="2025-11-26 06:27:42.678603346 +0000 UTC m=+146.039907298" Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.683981 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.183956654 +0000 UTC m=+146.545260606 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.684446 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" event={"ID":"e1cb46aa-dd22-49d0-9a14-dab875c02723","Type":"ContainerStarted","Data":"6a923baa7699dc01442480cc5cbaa7bc6be02821b91f5339ae4a78e34b810bae"} Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.700338 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" podStartSLOduration=124.700320696 podStartE2EDuration="2m4.700320696s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.700056849 +0000 UTC m=+146.061360821" watchObservedRunningTime="2025-11-26 06:27:42.700320696 +0000 UTC m=+146.061624648" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.710536 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kn6ts" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.724592 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-rb4zf" podStartSLOduration=124.724576845 podStartE2EDuration="2m4.724576845s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.722727104 +0000 UTC m=+146.084031056" watchObservedRunningTime="2025-11-26 06:27:42.724576845 +0000 UTC m=+146.085880787" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.779007 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2lfgm" podStartSLOduration=125.778989717 podStartE2EDuration="2m5.778989717s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.753233006 +0000 UTC m=+146.114536958" watchObservedRunningTime="2025-11-26 06:27:42.778989717 +0000 UTC m=+146.140293669" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.779662 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nmzpp" podStartSLOduration=124.779656076 podStartE2EDuration="2m4.779656076s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.778828373 +0000 UTC m=+146.140132325" watchObservedRunningTime="2025-11-26 06:27:42.779656076 +0000 UTC m=+146.140960028" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.780539 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.781062 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.281047364 +0000 UTC m=+146.642351326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.868696 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8lbfh" podStartSLOduration=125.868680523 podStartE2EDuration="2m5.868680523s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.864077076 +0000 UTC m=+146.225381078" watchObservedRunningTime="2025-11-26 06:27:42.868680523 +0000 UTC m=+146.229984475" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.881165 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.881393 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.381377083 +0000 UTC m=+146.742681035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.892029 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-49tww" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.929666 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vmb" podStartSLOduration=124.929652855 podStartE2EDuration="2m4.929652855s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.892170881 +0000 UTC m=+146.253474833" watchObservedRunningTime="2025-11-26 06:27:42.929652855 +0000 UTC m=+146.290956807" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.930306 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-49tww" podStartSLOduration=125.930301463 podStartE2EDuration="2m5.930301463s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.928267077 +0000 UTC m=+146.289571029" watchObservedRunningTime="2025-11-26 06:27:42.930301463 +0000 UTC m=+146.291605415" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.956570 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" podStartSLOduration=125.956553858 podStartE2EDuration="2m5.956553858s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.950882982 +0000 UTC m=+146.312186934" watchObservedRunningTime="2025-11-26 06:27:42.956553858 +0000 UTC m=+146.317857810" Nov 26 06:27:42 crc kubenswrapper[4775]: I1126 06:27:42.983398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:42 crc kubenswrapper[4775]: E1126 06:27:42.983727 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.483700117 +0000 UTC m=+146.845004069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.013406 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" podStartSLOduration=126.013387237 podStartE2EDuration="2m6.013387237s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:43.007972257 +0000 UTC m=+146.369276209" watchObservedRunningTime="2025-11-26 06:27:43.013387237 +0000 UTC m=+146.374691189" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.013560 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" podStartSLOduration=125.013557211 podStartE2EDuration="2m5.013557211s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:42.977910937 +0000 UTC m=+146.339214879" watchObservedRunningTime="2025-11-26 06:27:43.013557211 +0000 UTC m=+146.374861163" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.085517 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.085757 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.585743924 +0000 UTC m=+146.947047876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.186494 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.186826 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.686814733 +0000 UTC m=+147.048118685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.290534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.290792 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.790774653 +0000 UTC m=+147.152078605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.391569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.391960 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.891942835 +0000 UTC m=+147.253246787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.410640 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:43 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:43 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:43 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.410692 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.492889 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.493070 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.993043115 +0000 UTC m=+147.354347067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.493198 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.493707 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:43.993695893 +0000 UTC m=+147.354999845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.596377 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.597574 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.09754768 +0000 UTC m=+147.458851632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.597710 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.598181 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.098169787 +0000 UTC m=+147.459473739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.699698 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.700051 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.200037278 +0000 UTC m=+147.561341230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.737005 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9czjc" event={"ID":"cc93554a-637c-4af3-a239-bdfab76f780f","Type":"ContainerStarted","Data":"50c3c05bd91e32bca8a53f440c1dca7524d1eafaa6fa2433332977c1a9534e16"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.758728 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" event={"ID":"f1365993-a452-4e00-92e0-64e1c08ce1c2","Type":"ContainerStarted","Data":"90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.759385 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-z59ck container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.759451 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.766515 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" event={"ID":"6b0fb0ce-6df9-4e4c-8d48-b35699664601","Type":"ContainerStarted","Data":"498f1eeaa668a69668a9fdf89872c045159c0bc8d796f6305b5c607dac12eefb"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.766692 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.767995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rp4jq" event={"ID":"cfb25e56-d6ab-4e49-86de-9a20ba2ee02a","Type":"ContainerStarted","Data":"fa9d184142f46ac17b66c10b7e7db1bb6051a1d4ee82928e5d0529cc07c5ebe6"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.783014 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-zbzr7" event={"ID":"dd1a5521-45f9-4377-970e-4710f6749fa5","Type":"ContainerStarted","Data":"f39a2474ced57afc9335c167d6f0dcf8840750b7a3f689db3cdc91f9dd079031"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.789469 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" event={"ID":"764b5255-c5f2-4e67-8ad9-0050044df158","Type":"ContainerStarted","Data":"a4eccb1085c094fb62dfd469a384a0eb0401794988fa1470b0b46f9813410a85"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.797886 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kclct" event={"ID":"e0344f36-af1e-44ad-9f72-7a616d6e222c","Type":"ContainerStarted","Data":"16c3219b0fb82abda845c5f1f3702356b122bd5c53fc323b3252877daeec1d70"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.801824 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.802120 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.302108736 +0000 UTC m=+147.663412688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.803548 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9cq9z" event={"ID":"a11b6fbe-aabd-45e3-9302-3b3678bba644","Type":"ContainerStarted","Data":"3ff5142a09e8fdf22f8b883cd5eeffc9070be81176340717f0c63335d3fc4342"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.803596 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-9cq9z" event={"ID":"a11b6fbe-aabd-45e3-9302-3b3678bba644","Type":"ContainerStarted","Data":"179ba79d6c7175b7e9ed1b599654bb98f7745fa53ef37e7a10e145d560b2df35"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.803701 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.821870 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" podStartSLOduration=125.821855281 podStartE2EDuration="2m5.821855281s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:43.821147611 +0000 UTC m=+147.182451573" watchObservedRunningTime="2025-11-26 06:27:43.821855281 +0000 UTC m=+147.183159223" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.829949 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" event={"ID":"f1c14a59-51f9-475d-8867-aefb952b66dc","Type":"ContainerStarted","Data":"6fc71c90701bfb310cc54528c5c7452e79e8671c56661ccae3df4898ec05806a"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.829996 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" event={"ID":"f1c14a59-51f9-475d-8867-aefb952b66dc","Type":"ContainerStarted","Data":"b6b9b12aee3e89b166355d5429a9163092ea63a7edf96ef61c9f46f4584122d6"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.860236 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" event={"ID":"8a969da6-616f-4f99-8271-5a0d6a39b06c","Type":"ContainerStarted","Data":"61f48e90b1d042683950b1c1c2cad4c95604847e631a17eea50fe48e8bda9363"} Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.871116 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-db5pc" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.874924 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rp4jq" podStartSLOduration=7.874911495 podStartE2EDuration="7.874911495s" podCreationTimestamp="2025-11-26 06:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:43.852389123 +0000 UTC m=+147.213693075" watchObservedRunningTime="2025-11-26 06:27:43.874911495 +0000 UTC m=+147.236215447" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.877620 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-9cq9z" podStartSLOduration=7.877599019 podStartE2EDuration="7.877599019s" podCreationTimestamp="2025-11-26 06:27:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:43.874260737 +0000 UTC m=+147.235564679" watchObservedRunningTime="2025-11-26 06:27:43.877599019 +0000 UTC m=+147.238902971" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.898991 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-gt6r8" podStartSLOduration=125.898975409 podStartE2EDuration="2m5.898975409s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:43.898303361 +0000 UTC m=+147.259607313" watchObservedRunningTime="2025-11-26 06:27:43.898975409 +0000 UTC m=+147.260279351" Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.905372 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:43 crc kubenswrapper[4775]: E1126 06:27:43.905689 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.405674134 +0000 UTC m=+147.766978076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:43 crc kubenswrapper[4775]: I1126 06:27:43.925652 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-t29hf" podStartSLOduration=125.925637465 podStartE2EDuration="2m5.925637465s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:43.925133381 +0000 UTC m=+147.286437343" watchObservedRunningTime="2025-11-26 06:27:43.925637465 +0000 UTC m=+147.286941417" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.009224 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.009669 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.509652323 +0000 UTC m=+147.870956275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.114452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.114619 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.614596899 +0000 UTC m=+147.975900851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.114800 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.115087 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.615079733 +0000 UTC m=+147.976383685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.215607 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.215972 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.715958417 +0000 UTC m=+148.077262369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.228624 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kz9zl" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.317531 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.318145 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.818132087 +0000 UTC m=+148.179436039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.376180 4775 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.405324 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7bwzr"] Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.408495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.413144 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8dhs8" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.419757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.419898 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fbbn\" (UniqueName: \"kubernetes.io/projected/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-kube-api-access-8fbbn\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.419939 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-catalog-content\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.419988 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-utilities\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.420089 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 06:27:44.920074291 +0000 UTC m=+148.281378243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.421176 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:44 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:44 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:44 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.421207 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.424495 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7bwzr"] Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.426196 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.511099 4775 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-26T06:27:44.376225071Z","Handler":null,"Name":""} Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.520903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-catalog-content\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.520977 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-utilities\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.521004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.521049 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fbbn\" (UniqueName: \"kubernetes.io/projected/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-kube-api-access-8fbbn\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.521363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-catalog-content\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.521378 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-utilities\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: E1126 06:27:44.521403 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 06:27:45.021388497 +0000 UTC m=+148.382692449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-4h8zh" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.525674 4775 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.525727 4775 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.552054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fbbn\" (UniqueName: \"kubernetes.io/projected/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-kube-api-access-8fbbn\") pod \"certified-operators-7bwzr\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.596724 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9p8qj"] Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.597577 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.598907 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.607955 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9p8qj"] Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.621502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.621756 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qfb6\" (UniqueName: \"kubernetes.io/projected/4bdf9960-b1db-4086-9492-1b656cbe6e1b-kube-api-access-4qfb6\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.621848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-catalog-content\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.621876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-utilities\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.657848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.723447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-catalog-content\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.723510 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-utilities\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.723544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.723569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qfb6\" (UniqueName: \"kubernetes.io/projected/4bdf9960-b1db-4086-9492-1b656cbe6e1b-kube-api-access-4qfb6\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.725015 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-utilities\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.725177 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-catalog-content\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.726444 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.726900 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.726935 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.744812 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qfb6\" (UniqueName: \"kubernetes.io/projected/4bdf9960-b1db-4086-9492-1b656cbe6e1b-kube-api-access-4qfb6\") pod \"community-operators-9p8qj\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.789507 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nk6kf"] Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.790590 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.801150 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk6kf"] Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.823691 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-4h8zh\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.824271 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-catalog-content\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.824329 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88c99\" (UniqueName: \"kubernetes.io/projected/810db382-46fe-4ccc-b05e-5e6f8f540f8e-kube-api-access-88c99\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.824586 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-utilities\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.893420 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" event={"ID":"764b5255-c5f2-4e67-8ad9-0050044df158","Type":"ContainerStarted","Data":"5a76eb13b944c43cccd3b3c316c4e9678918112eb1573e57a524d2d9c9cc319e"} Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.893556 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" event={"ID":"764b5255-c5f2-4e67-8ad9-0050044df158","Type":"ContainerStarted","Data":"b10c5834766490b03fcc82fc958cb796001850f1a14097e758e6a6e693dc0aef"} Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.909991 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.927828 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pv8cb" podStartSLOduration=9.927814485 podStartE2EDuration="9.927814485s" podCreationTimestamp="2025-11-26 06:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:44.925008947 +0000 UTC m=+148.286312899" watchObservedRunningTime="2025-11-26 06:27:44.927814485 +0000 UTC m=+148.289118437" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.937066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-utilities\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.937276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-catalog-content\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.937320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88c99\" (UniqueName: \"kubernetes.io/projected/810db382-46fe-4ccc-b05e-5e6f8f540f8e-kube-api-access-88c99\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.938187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-utilities\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.942443 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-catalog-content\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.951923 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:27:44 crc kubenswrapper[4775]: I1126 06:27:44.966545 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88c99\" (UniqueName: \"kubernetes.io/projected/810db382-46fe-4ccc-b05e-5e6f8f540f8e-kube-api-access-88c99\") pod \"certified-operators-nk6kf\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.004902 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hx9c6"] Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.006734 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.011055 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hx9c6"] Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.026085 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.038435 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bp42\" (UniqueName: \"kubernetes.io/projected/69d43e02-199e-48d1-a4a0-a50248a3946e-kube-api-access-6bp42\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.038662 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-utilities\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.038696 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-catalog-content\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.109942 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.139677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.139741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bp42\" (UniqueName: \"kubernetes.io/projected/69d43e02-199e-48d1-a4a0-a50248a3946e-kube-api-access-6bp42\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.139770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-utilities\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.139793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-catalog-content\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.139815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.140732 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.142080 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-utilities\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.142516 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-catalog-content\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.144865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.149689 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9p8qj"] Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.158648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bp42\" (UniqueName: \"kubernetes.io/projected/69d43e02-199e-48d1-a4a0-a50248a3946e-kube-api-access-6bp42\") pod \"community-operators-hx9c6\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.174400 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7bwzr"] Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.241434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.241493 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.243973 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4h8zh"] Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.245056 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.245088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:45 crc kubenswrapper[4775]: W1126 06:27:45.267556 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod088e2d8e_dfb4_4a5d_b681_d7d46573b629.slice/crio-33c7405d32376be90f51708bac779f6618da214bd757b0f905e3bf5ac9a9fead WatchSource:0}: Error finding container 33c7405d32376be90f51708bac779f6618da214bd757b0f905e3bf5ac9a9fead: Status 404 returned error can't find the container with id 33c7405d32376be90f51708bac779f6618da214bd757b0f905e3bf5ac9a9fead Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.333023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.339546 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.352287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.359692 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk6kf"] Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.378234 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 06:27:45 crc kubenswrapper[4775]: W1126 06:27:45.378899 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod810db382_46fe_4ccc_b05e_5e6f8f540f8e.slice/crio-0cadb339686dadcd0a2f561a95f5b9c5d5fc5f292c58ee8ad38348f93c6242e8 WatchSource:0}: Error finding container 0cadb339686dadcd0a2f561a95f5b9c5d5fc5f292c58ee8ad38348f93c6242e8: Status 404 returned error can't find the container with id 0cadb339686dadcd0a2f561a95f5b9c5d5fc5f292c58ee8ad38348f93c6242e8 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.392126 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.418144 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:45 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:45 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:45 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.418208 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.551951 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hx9c6"] Nov 26 06:27:45 crc kubenswrapper[4775]: W1126 06:27:45.567027 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69d43e02_199e_48d1_a4a0_a50248a3946e.slice/crio-46a8d77f522cd4fabf71c903b780b5d47f09c6d1759114493be3bc6406b4a695 WatchSource:0}: Error finding container 46a8d77f522cd4fabf71c903b780b5d47f09c6d1759114493be3bc6406b4a695: Status 404 returned error can't find the container with id 46a8d77f522cd4fabf71c903b780b5d47f09c6d1759114493be3bc6406b4a695 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.910545 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4ad928b68cd6b0eec779b228c63c9e77ffdc387d05c06108639d5b1f0b8b8916"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.912822 4775 generic.go:334] "Generic (PLEG): container finished" podID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerID="2533b128784785dadcd2bee676b6cfa347ab8a2061b2c80f18d1b9a933db53e8" exitCode=0 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.912887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx9c6" event={"ID":"69d43e02-199e-48d1-a4a0-a50248a3946e","Type":"ContainerDied","Data":"2533b128784785dadcd2bee676b6cfa347ab8a2061b2c80f18d1b9a933db53e8"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.912913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx9c6" event={"ID":"69d43e02-199e-48d1-a4a0-a50248a3946e","Type":"ContainerStarted","Data":"46a8d77f522cd4fabf71c903b780b5d47f09c6d1759114493be3bc6406b4a695"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.914406 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.914980 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerID="64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd" exitCode=0 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.915027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bwzr" event={"ID":"7ea68cee-a3d6-4ff7-b668-d87599fa4efd","Type":"ContainerDied","Data":"64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.915042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bwzr" event={"ID":"7ea68cee-a3d6-4ff7-b668-d87599fa4efd","Type":"ContainerStarted","Data":"6134b8cb59ee28ccc393b5fa15e7597ec74d6a6a50c2eebbd117bd4de9fac7b1"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.916931 4775 generic.go:334] "Generic (PLEG): container finished" podID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerID="7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd" exitCode=0 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.916986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6kf" event={"ID":"810db382-46fe-4ccc-b05e-5e6f8f540f8e","Type":"ContainerDied","Data":"7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.917010 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6kf" event={"ID":"810db382-46fe-4ccc-b05e-5e6f8f540f8e","Type":"ContainerStarted","Data":"0cadb339686dadcd0a2f561a95f5b9c5d5fc5f292c58ee8ad38348f93c6242e8"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.920477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" event={"ID":"088e2d8e-dfb4-4a5d-b681-d7d46573b629","Type":"ContainerStarted","Data":"216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.920526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" event={"ID":"088e2d8e-dfb4-4a5d-b681-d7d46573b629","Type":"ContainerStarted","Data":"33c7405d32376be90f51708bac779f6618da214bd757b0f905e3bf5ac9a9fead"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.920597 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.922520 4775 generic.go:334] "Generic (PLEG): container finished" podID="ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" containerID="d95be0bed38af2377ee70712b06f96a07c0956b6629e4603ad5aeee39865f917" exitCode=0 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.922588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" event={"ID":"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3","Type":"ContainerDied","Data":"d95be0bed38af2377ee70712b06f96a07c0956b6629e4603ad5aeee39865f917"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.923841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"726231c9c21aa6a8e573ba951d665c63397a4ba21bd6d34f13e7ac533942d946"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.925048 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e9822fda3749ad5025bab8f6300d20fc1862a1c4120d4719e2df466ad14b1058"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.929768 4775 generic.go:334] "Generic (PLEG): container finished" podID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerID="446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59" exitCode=0 Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.930181 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p8qj" event={"ID":"4bdf9960-b1db-4086-9492-1b656cbe6e1b","Type":"ContainerDied","Data":"446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59"} Nov 26 06:27:45 crc kubenswrapper[4775]: I1126 06:27:45.930244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p8qj" event={"ID":"4bdf9960-b1db-4086-9492-1b656cbe6e1b","Type":"ContainerStarted","Data":"1545686d3e732e381cd24ae74bd3fd250837d3d115bb5afce042439223d7b43e"} Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.016763 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" podStartSLOduration=128.01674668 podStartE2EDuration="2m8.01674668s" podCreationTimestamp="2025-11-26 06:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:46.015238968 +0000 UTC m=+149.376542930" watchObservedRunningTime="2025-11-26 06:27:46.01674668 +0000 UTC m=+149.378050632" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.395002 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hpsd2"] Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.396626 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.398579 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.407451 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hpsd2"] Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.409116 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:46 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:46 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:46 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.409178 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.428977 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.429865 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.431500 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.432107 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.464641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95788dae-18a5-49ee-9f1c-8e73527f9673-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.464691 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kxrj\" (UniqueName: \"kubernetes.io/projected/e21090b1-0905-474e-9148-fd349044b039-kube-api-access-7kxrj\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.464742 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-utilities\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.464778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-catalog-content\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.464915 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95788dae-18a5-49ee-9f1c-8e73527f9673-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.478698 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-catalog-content\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95788dae-18a5-49ee-9f1c-8e73527f9673-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567305 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95788dae-18a5-49ee-9f1c-8e73527f9673-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kxrj\" (UniqueName: \"kubernetes.io/projected/e21090b1-0905-474e-9148-fd349044b039-kube-api-access-7kxrj\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-utilities\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567388 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95788dae-18a5-49ee-9f1c-8e73527f9673-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567590 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-catalog-content\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.567949 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-utilities\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.594393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95788dae-18a5-49ee-9f1c-8e73527f9673-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.594610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kxrj\" (UniqueName: \"kubernetes.io/projected/e21090b1-0905-474e-9148-fd349044b039-kube-api-access-7kxrj\") pod \"redhat-marketplace-hpsd2\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.753034 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.762424 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.803130 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5rt8d"] Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.805015 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.812041 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rt8d"] Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.871837 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-catalog-content\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.871935 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-utilities\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.872001 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95njv\" (UniqueName: \"kubernetes.io/projected/eaf163ed-6550-4efb-b83e-8fd80bea0935-kube-api-access-95njv\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.951395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1b2e302ee98d9eb9b684f50286b8c034f9334e5d7de68642b45e8731c81df3da"} Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.952488 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.954346 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e8e4b14bafb7bacce2026cc195322249ff4f6e7db11aef1227af9b537b31f9f3"} Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.956986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c7d1f44b420c7ffea4baf24739ab0e185e80bab27ac812e901de052f2a8d26cd"} Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.973030 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-catalog-content\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.973093 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-utilities\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.973137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95njv\" (UniqueName: \"kubernetes.io/projected/eaf163ed-6550-4efb-b83e-8fd80bea0935-kube-api-access-95njv\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.974161 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-utilities\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:46 crc kubenswrapper[4775]: I1126 06:27:46.974343 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-catalog-content\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.003701 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95njv\" (UniqueName: \"kubernetes.io/projected/eaf163ed-6550-4efb-b83e-8fd80bea0935-kube-api-access-95njv\") pod \"redhat-marketplace-5rt8d\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.097627 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.145968 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.239724 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.375970 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-44mw9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.376112 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-44mw9" podUID="2278df09-65b1-440e-a613-7bea0b8c73cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.376517 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-44mw9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.376652 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-44mw9" podUID="2278df09-65b1-440e-a613-7bea0b8c73cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.378932 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-config-volume\") pod \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.379027 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46jg9\" (UniqueName: \"kubernetes.io/projected/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-kube-api-access-46jg9\") pod \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.379154 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-secret-volume\") pod \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\" (UID: \"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3\") " Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.381820 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hpsd2"] Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.384357 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-config-volume" (OuterVolumeSpecName: "config-volume") pod "ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" (UID: "ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.390336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-kube-api-access-46jg9" (OuterVolumeSpecName: "kube-api-access-46jg9") pod "ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" (UID: "ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3"). InnerVolumeSpecName "kube-api-access-46jg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.391609 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" (UID: "ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.411176 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:47 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:47 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:47 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.411227 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.458583 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rt8d"] Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.480879 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.480905 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46jg9\" (UniqueName: \"kubernetes.io/projected/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-kube-api-access-46jg9\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.480914 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:47 crc kubenswrapper[4775]: W1126 06:27:47.553015 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode21090b1_0905_474e_9148_fd349044b039.slice/crio-1767fff75c505352c7aa2f8cdff6b0f5fef44e2000ced6fa49aa7fa5a14e9e0a WatchSource:0}: Error finding container 1767fff75c505352c7aa2f8cdff6b0f5fef44e2000ced6fa49aa7fa5a14e9e0a: Status 404 returned error can't find the container with id 1767fff75c505352c7aa2f8cdff6b0f5fef44e2000ced6fa49aa7fa5a14e9e0a Nov 26 06:27:47 crc kubenswrapper[4775]: W1126 06:27:47.557299 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaf163ed_6550_4efb_b83e_8fd80bea0935.slice/crio-296f753163ada7f2f4e6130686d1898889068ce43c17eae8d23216fdaadb7635 WatchSource:0}: Error finding container 296f753163ada7f2f4e6130686d1898889068ce43c17eae8d23216fdaadb7635: Status 404 returned error can't find the container with id 296f753163ada7f2f4e6130686d1898889068ce43c17eae8d23216fdaadb7635 Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.599017 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.605018 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kclct" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.797295 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sjc6t"] Nov 26 06:27:47 crc kubenswrapper[4775]: E1126 06:27:47.799575 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" containerName="collect-profiles" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.799643 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" containerName="collect-profiles" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.799801 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" containerName="collect-profiles" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.800638 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.812183 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.822554 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sjc6t"] Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.892937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-utilities\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.892980 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zc2z\" (UniqueName: \"kubernetes.io/projected/64dc0036-294e-4f2b-9b9d-cd540c42e558-kube-api-access-8zc2z\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.893014 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-catalog-content\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.972564 4775 generic.go:334] "Generic (PLEG): container finished" podID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerID="f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255" exitCode=0 Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.972678 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerDied","Data":"f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255"} Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.972771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerStarted","Data":"296f753163ada7f2f4e6130686d1898889068ce43c17eae8d23216fdaadb7635"} Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.975103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95788dae-18a5-49ee-9f1c-8e73527f9673","Type":"ContainerStarted","Data":"f910eda02531e58f61eb6abb253e7aaee61711c17eb6039f0eb8bd6dd29a839b"} Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.975167 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95788dae-18a5-49ee-9f1c-8e73527f9673","Type":"ContainerStarted","Data":"b793eebd72ffe6c44008bdbdab6eb51ad5abd92c64f45c3b305d1adab29b808e"} Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.977230 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" event={"ID":"ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3","Type":"ContainerDied","Data":"d3296732f9660e24bedf1d12c43b482380835856db689327eca81c33c2e3a16c"} Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.977275 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3296732f9660e24bedf1d12c43b482380835856db689327eca81c33c2e3a16c" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.977406 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.996160 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-utilities\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.996208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zc2z\" (UniqueName: \"kubernetes.io/projected/64dc0036-294e-4f2b-9b9d-cd540c42e558-kube-api-access-8zc2z\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.996262 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-catalog-content\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.998452 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-utilities\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.998596 4775 generic.go:334] "Generic (PLEG): container finished" podID="e21090b1-0905-474e-9148-fd349044b039" containerID="564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80" exitCode=0 Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.999073 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hpsd2" event={"ID":"e21090b1-0905-474e-9148-fd349044b039","Type":"ContainerDied","Data":"564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80"} Nov 26 06:27:47 crc kubenswrapper[4775]: I1126 06:27:47.999112 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hpsd2" event={"ID":"e21090b1-0905-474e-9148-fd349044b039","Type":"ContainerStarted","Data":"1767fff75c505352c7aa2f8cdff6b0f5fef44e2000ced6fa49aa7fa5a14e9e0a"} Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.000177 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-catalog-content\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.017747 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zc2z\" (UniqueName: \"kubernetes.io/projected/64dc0036-294e-4f2b-9b9d-cd540c42e558-kube-api-access-8zc2z\") pod \"redhat-operators-sjc6t\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.067987 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.067971763 podStartE2EDuration="2.067971763s" podCreationTimestamp="2025-11-26 06:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:27:48.065953938 +0000 UTC m=+151.427257890" watchObservedRunningTime="2025-11-26 06:27:48.067971763 +0000 UTC m=+151.429275715" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.195542 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rj2bg"] Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.196601 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.210620 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rj2bg"] Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.211192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.278096 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.278332 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.279252 4775 patch_prober.go:28] interesting pod/console-f9d7485db-vl6j8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.279317 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vl6j8" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.299225 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-catalog-content\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.299264 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq9wt\" (UniqueName: \"kubernetes.io/projected/f5dc821f-58e0-47ec-8aac-d67b333013a4-kube-api-access-wq9wt\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.299355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-utilities\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.400739 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-utilities\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.400875 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-catalog-content\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.400900 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq9wt\" (UniqueName: \"kubernetes.io/projected/f5dc821f-58e0-47ec-8aac-d67b333013a4-kube-api-access-wq9wt\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.401278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-utilities\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.401773 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-catalog-content\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.406893 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.410033 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:48 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:48 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:48 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.410063 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.417085 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq9wt\" (UniqueName: \"kubernetes.io/projected/f5dc821f-58e0-47ec-8aac-d67b333013a4-kube-api-access-wq9wt\") pod \"redhat-operators-rj2bg\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.473523 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sjc6t"] Nov 26 06:27:48 crc kubenswrapper[4775]: W1126 06:27:48.490289 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64dc0036_294e_4f2b_9b9d_cd540c42e558.slice/crio-b81c3b3f3fcbcb6c5e891e4c9ef3e469ae6b0344e5908e0041128b63f1295ddf WatchSource:0}: Error finding container b81c3b3f3fcbcb6c5e891e4c9ef3e469ae6b0344e5908e0041128b63f1295ddf: Status 404 returned error can't find the container with id b81c3b3f3fcbcb6c5e891e4c9ef3e469ae6b0344e5908e0041128b63f1295ddf Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.516201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.560232 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.560923 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.569303 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.569687 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.571523 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.704257 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.704325 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.805813 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.806247 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.806835 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.822193 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rj2bg"] Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.838029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:48 crc kubenswrapper[4775]: I1126 06:27:48.904578 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.013291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerStarted","Data":"6bedf77141fde9cb742b821814f34891a658dbfdffd398378d525a0546464261"} Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.017656 4775 generic.go:334] "Generic (PLEG): container finished" podID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerID="de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1" exitCode=0 Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.017699 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerDied","Data":"de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1"} Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.017728 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerStarted","Data":"b81c3b3f3fcbcb6c5e891e4c9ef3e469ae6b0344e5908e0041128b63f1295ddf"} Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.021694 4775 generic.go:334] "Generic (PLEG): container finished" podID="95788dae-18a5-49ee-9f1c-8e73527f9673" containerID="f910eda02531e58f61eb6abb253e7aaee61711c17eb6039f0eb8bd6dd29a839b" exitCode=0 Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.022046 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95788dae-18a5-49ee-9f1c-8e73527f9673","Type":"ContainerDied","Data":"f910eda02531e58f61eb6abb253e7aaee61711c17eb6039f0eb8bd6dd29a839b"} Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.256575 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 06:27:49 crc kubenswrapper[4775]: W1126 06:27:49.308340 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod23b2a7d9_a51e_48a3_bbec_3bfb3dc5ae36.slice/crio-1df0723b8abf8e26be5cd666b14b8d12633bde2add59d0f0462de84f0d7a6fc0 WatchSource:0}: Error finding container 1df0723b8abf8e26be5cd666b14b8d12633bde2add59d0f0462de84f0d7a6fc0: Status 404 returned error can't find the container with id 1df0723b8abf8e26be5cd666b14b8d12633bde2add59d0f0462de84f0d7a6fc0 Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.409906 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:49 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:49 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:49 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:49 crc kubenswrapper[4775]: I1126 06:27:49.409960 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.051018 4775 generic.go:334] "Generic (PLEG): container finished" podID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerID="c73c84d3cff170e8666f0d22dcb395b56c19004f652c7d001ede6a8e488e12ee" exitCode=0 Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.051174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerDied","Data":"c73c84d3cff170e8666f0d22dcb395b56c19004f652c7d001ede6a8e488e12ee"} Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.059638 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36","Type":"ContainerStarted","Data":"1df0723b8abf8e26be5cd666b14b8d12633bde2add59d0f0462de84f0d7a6fc0"} Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.420393 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:50 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:50 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:50 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.420440 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.434292 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.548758 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95788dae-18a5-49ee-9f1c-8e73527f9673-kubelet-dir\") pod \"95788dae-18a5-49ee-9f1c-8e73527f9673\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.548842 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95788dae-18a5-49ee-9f1c-8e73527f9673-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "95788dae-18a5-49ee-9f1c-8e73527f9673" (UID: "95788dae-18a5-49ee-9f1c-8e73527f9673"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.548871 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95788dae-18a5-49ee-9f1c-8e73527f9673-kube-api-access\") pod \"95788dae-18a5-49ee-9f1c-8e73527f9673\" (UID: \"95788dae-18a5-49ee-9f1c-8e73527f9673\") " Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.549835 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/95788dae-18a5-49ee-9f1c-8e73527f9673-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.575437 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95788dae-18a5-49ee-9f1c-8e73527f9673-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "95788dae-18a5-49ee-9f1c-8e73527f9673" (UID: "95788dae-18a5-49ee-9f1c-8e73527f9673"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:27:50 crc kubenswrapper[4775]: I1126 06:27:50.651143 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95788dae-18a5-49ee-9f1c-8e73527f9673-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.124588 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.126811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"95788dae-18a5-49ee-9f1c-8e73527f9673","Type":"ContainerDied","Data":"b793eebd72ffe6c44008bdbdab6eb51ad5abd92c64f45c3b305d1adab29b808e"} Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.126852 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b793eebd72ffe6c44008bdbdab6eb51ad5abd92c64f45c3b305d1adab29b808e" Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.169862 4775 generic.go:334] "Generic (PLEG): container finished" podID="23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36" containerID="afbaabf407404719cd30defc86f75ea1f60c561e68482e43b014f9fbf25f17d4" exitCode=0 Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.169900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36","Type":"ContainerDied","Data":"afbaabf407404719cd30defc86f75ea1f60c561e68482e43b014f9fbf25f17d4"} Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.409011 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:51 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:51 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:51 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.409073 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.420010 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:27:51 crc kubenswrapper[4775]: I1126 06:27:51.420064 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:27:52 crc kubenswrapper[4775]: I1126 06:27:52.409192 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:52 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:52 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:52 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:52 crc kubenswrapper[4775]: I1126 06:27:52.409453 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:53 crc kubenswrapper[4775]: I1126 06:27:53.408215 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:53 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:53 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:53 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:53 crc kubenswrapper[4775]: I1126 06:27:53.408283 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:54 crc kubenswrapper[4775]: I1126 06:27:54.232477 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-9cq9z" Nov 26 06:27:54 crc kubenswrapper[4775]: I1126 06:27:54.407759 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:54 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:54 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:54 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:54 crc kubenswrapper[4775]: I1126 06:27:54.407804 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:55 crc kubenswrapper[4775]: I1126 06:27:55.409169 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:55 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:55 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:55 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:55 crc kubenswrapper[4775]: I1126 06:27:55.409441 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:56 crc kubenswrapper[4775]: I1126 06:27:56.408178 4775 patch_prober.go:28] interesting pod/router-default-5444994796-drckl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 06:27:56 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Nov 26 06:27:56 crc kubenswrapper[4775]: [+]process-running ok Nov 26 06:27:56 crc kubenswrapper[4775]: healthz check failed Nov 26 06:27:56 crc kubenswrapper[4775]: I1126 06:27:56.408229 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-drckl" podUID="f02e8337-4b5c-4ec3-b0a2-38d674cbfe54" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.399073 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-44mw9" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.432077 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.435369 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-drckl" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.475281 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.593457 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kube-api-access\") pod \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.593569 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kubelet-dir\") pod \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\" (UID: \"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36\") " Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.593810 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36" (UID: "23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.594046 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.600487 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36" (UID: "23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:27:57 crc kubenswrapper[4775]: I1126 06:27:57.695059 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:27:58 crc kubenswrapper[4775]: I1126 06:27:58.211795 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36","Type":"ContainerDied","Data":"1df0723b8abf8e26be5cd666b14b8d12633bde2add59d0f0462de84f0d7a6fc0"} Nov 26 06:27:58 crc kubenswrapper[4775]: I1126 06:27:58.211836 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1df0723b8abf8e26be5cd666b14b8d12633bde2add59d0f0462de84f0d7a6fc0" Nov 26 06:27:58 crc kubenswrapper[4775]: I1126 06:27:58.211810 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 06:27:58 crc kubenswrapper[4775]: I1126 06:27:58.277709 4775 patch_prober.go:28] interesting pod/console-f9d7485db-vl6j8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Nov 26 06:27:58 crc kubenswrapper[4775]: I1126 06:27:58.277796 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vl6j8" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Nov 26 06:28:00 crc kubenswrapper[4775]: I1126 06:28:00.334219 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:28:00 crc kubenswrapper[4775]: I1126 06:28:00.342819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e24b0f5-8555-40a4-a965-e67d5cd305a6-metrics-certs\") pod \"network-metrics-daemon-n98r8\" (UID: \"1e24b0f5-8555-40a4-a965-e67d5cd305a6\") " pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:28:00 crc kubenswrapper[4775]: I1126 06:28:00.403439 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n98r8" Nov 26 06:28:05 crc kubenswrapper[4775]: I1126 06:28:05.033485 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:28:08 crc kubenswrapper[4775]: I1126 06:28:08.314237 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:28:08 crc kubenswrapper[4775]: I1126 06:28:08.319636 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:28:10 crc kubenswrapper[4775]: E1126 06:28:10.378309 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 26 06:28:10 crc kubenswrapper[4775]: E1126 06:28:10.379701 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7kxrj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-hpsd2_openshift-marketplace(e21090b1-0905-474e-9148-fd349044b039): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 06:28:10 crc kubenswrapper[4775]: E1126 06:28:10.381086 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-hpsd2" podUID="e21090b1-0905-474e-9148-fd349044b039" Nov 26 06:28:13 crc kubenswrapper[4775]: E1126 06:28:13.051773 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-hpsd2" podUID="e21090b1-0905-474e-9148-fd349044b039" Nov 26 06:28:13 crc kubenswrapper[4775]: E1126 06:28:13.519117 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 26 06:28:13 crc kubenswrapper[4775]: E1126 06:28:13.519588 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-95njv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5rt8d_openshift-marketplace(eaf163ed-6550-4efb-b83e-8fd80bea0935): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 06:28:13 crc kubenswrapper[4775]: E1126 06:28:13.520820 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5rt8d" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" Nov 26 06:28:13 crc kubenswrapper[4775]: I1126 06:28:13.621451 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n98r8"] Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.286877 4775 generic.go:334] "Generic (PLEG): container finished" podID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerID="176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f" exitCode=0 Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.286932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p8qj" event={"ID":"4bdf9960-b1db-4086-9492-1b656cbe6e1b","Type":"ContainerDied","Data":"176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.289845 4775 generic.go:334] "Generic (PLEG): container finished" podID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerID="70c16714ab9a989226bcd62795d5eae90c43cd6d86531494ba464912f104be0c" exitCode=0 Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.289915 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx9c6" event={"ID":"69d43e02-199e-48d1-a4a0-a50248a3946e","Type":"ContainerDied","Data":"70c16714ab9a989226bcd62795d5eae90c43cd6d86531494ba464912f104be0c"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.292288 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerID="18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a" exitCode=0 Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.292361 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bwzr" event={"ID":"7ea68cee-a3d6-4ff7-b668-d87599fa4efd","Type":"ContainerDied","Data":"18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.297383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerStarted","Data":"984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.299065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n98r8" event={"ID":"1e24b0f5-8555-40a4-a965-e67d5cd305a6","Type":"ContainerStarted","Data":"dba202b4d489fc36bd13c8fc01bf9dddf555c90db7b6a37a0c958a7b937c0052"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.299102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n98r8" event={"ID":"1e24b0f5-8555-40a4-a965-e67d5cd305a6","Type":"ContainerStarted","Data":"36d3f9f2939203ebec6168de1699a5f58c1f2698d9cffb6b245d24915d417980"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.302732 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerStarted","Data":"edbb9e673d0832dd0cc4bbf4a0cf4d025acaf0fd9e247dfdec5fde2d46d042a1"} Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.313183 4775 generic.go:334] "Generic (PLEG): container finished" podID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerID="21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1" exitCode=0 Nov 26 06:28:14 crc kubenswrapper[4775]: I1126 06:28:14.314319 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6kf" event={"ID":"810db382-46fe-4ccc-b05e-5e6f8f540f8e","Type":"ContainerDied","Data":"21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1"} Nov 26 06:28:14 crc kubenswrapper[4775]: E1126 06:28:14.315238 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5rt8d" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" Nov 26 06:28:15 crc kubenswrapper[4775]: I1126 06:28:15.320327 4775 generic.go:334] "Generic (PLEG): container finished" podID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerID="984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa" exitCode=0 Nov 26 06:28:15 crc kubenswrapper[4775]: I1126 06:28:15.320413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerDied","Data":"984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa"} Nov 26 06:28:15 crc kubenswrapper[4775]: I1126 06:28:15.322969 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n98r8" event={"ID":"1e24b0f5-8555-40a4-a965-e67d5cd305a6","Type":"ContainerStarted","Data":"9285da668c75519957febfa1158fcfbdd1de8220394edb0a7b531d3a9c9e996d"} Nov 26 06:28:15 crc kubenswrapper[4775]: I1126 06:28:15.325179 4775 generic.go:334] "Generic (PLEG): container finished" podID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerID="edbb9e673d0832dd0cc4bbf4a0cf4d025acaf0fd9e247dfdec5fde2d46d042a1" exitCode=0 Nov 26 06:28:15 crc kubenswrapper[4775]: I1126 06:28:15.325215 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerDied","Data":"edbb9e673d0832dd0cc4bbf4a0cf4d025acaf0fd9e247dfdec5fde2d46d042a1"} Nov 26 06:28:15 crc kubenswrapper[4775]: I1126 06:28:15.363440 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-n98r8" podStartSLOduration=158.363419203 podStartE2EDuration="2m38.363419203s" podCreationTimestamp="2025-11-26 06:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:28:15.360932064 +0000 UTC m=+178.722236026" watchObservedRunningTime="2025-11-26 06:28:15.363419203 +0000 UTC m=+178.724723165" Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.331908 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerStarted","Data":"a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36"} Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.335216 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerStarted","Data":"a9523fcc45fcd882d1780521daf6d31b4eea97212430d7185c77443d1afb5db1"} Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.337501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6kf" event={"ID":"810db382-46fe-4ccc-b05e-5e6f8f540f8e","Type":"ContainerStarted","Data":"741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77"} Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.339357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p8qj" event={"ID":"4bdf9960-b1db-4086-9492-1b656cbe6e1b","Type":"ContainerStarted","Data":"34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e"} Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.341230 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx9c6" event={"ID":"69d43e02-199e-48d1-a4a0-a50248a3946e","Type":"ContainerStarted","Data":"0645a504dcc3b11fc648460913b693f3c8c441bad0bf3c8ac9417fd09e8b67ee"} Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.343794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bwzr" event={"ID":"7ea68cee-a3d6-4ff7-b668-d87599fa4efd","Type":"ContainerStarted","Data":"aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e"} Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.350188 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sjc6t" podStartSLOduration=2.523353822 podStartE2EDuration="29.350176877s" podCreationTimestamp="2025-11-26 06:27:47 +0000 UTC" firstStartedPulling="2025-11-26 06:27:49.018949311 +0000 UTC m=+152.380253253" lastFinishedPulling="2025-11-26 06:28:15.845772346 +0000 UTC m=+179.207076308" observedRunningTime="2025-11-26 06:28:16.34884267 +0000 UTC m=+179.710146632" watchObservedRunningTime="2025-11-26 06:28:16.350176877 +0000 UTC m=+179.711480829" Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.388199 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hx9c6" podStartSLOduration=2.857896395 podStartE2EDuration="32.388183186s" podCreationTimestamp="2025-11-26 06:27:44 +0000 UTC" firstStartedPulling="2025-11-26 06:27:45.914168948 +0000 UTC m=+149.275472900" lastFinishedPulling="2025-11-26 06:28:15.444455719 +0000 UTC m=+178.805759691" observedRunningTime="2025-11-26 06:28:16.386490319 +0000 UTC m=+179.747794291" watchObservedRunningTime="2025-11-26 06:28:16.388183186 +0000 UTC m=+179.749487138" Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.388601 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rj2bg" podStartSLOduration=2.465410222 podStartE2EDuration="28.388596917s" podCreationTimestamp="2025-11-26 06:27:48 +0000 UTC" firstStartedPulling="2025-11-26 06:27:50.052801025 +0000 UTC m=+153.414104977" lastFinishedPulling="2025-11-26 06:28:15.97598772 +0000 UTC m=+179.337291672" observedRunningTime="2025-11-26 06:28:16.368259596 +0000 UTC m=+179.729563568" watchObservedRunningTime="2025-11-26 06:28:16.388596917 +0000 UTC m=+179.749900869" Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.419102 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9p8qj" podStartSLOduration=3.376961181 podStartE2EDuration="32.419084578s" podCreationTimestamp="2025-11-26 06:27:44 +0000 UTC" firstStartedPulling="2025-11-26 06:27:45.932381461 +0000 UTC m=+149.293685413" lastFinishedPulling="2025-11-26 06:28:14.974504828 +0000 UTC m=+178.335808810" observedRunningTime="2025-11-26 06:28:16.418157753 +0000 UTC m=+179.779461715" watchObservedRunningTime="2025-11-26 06:28:16.419084578 +0000 UTC m=+179.780388530" Nov 26 06:28:16 crc kubenswrapper[4775]: I1126 06:28:16.420950 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nk6kf" podStartSLOduration=2.867845799 podStartE2EDuration="32.42094271s" podCreationTimestamp="2025-11-26 06:27:44 +0000 UTC" firstStartedPulling="2025-11-26 06:27:45.91822643 +0000 UTC m=+149.279530382" lastFinishedPulling="2025-11-26 06:28:15.471323341 +0000 UTC m=+178.832627293" observedRunningTime="2025-11-26 06:28:16.402305555 +0000 UTC m=+179.763609517" watchObservedRunningTime="2025-11-26 06:28:16.42094271 +0000 UTC m=+179.782246662" Nov 26 06:28:18 crc kubenswrapper[4775]: I1126 06:28:18.212642 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:28:18 crc kubenswrapper[4775]: I1126 06:28:18.213155 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:28:18 crc kubenswrapper[4775]: I1126 06:28:18.516878 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:28:18 crc kubenswrapper[4775]: I1126 06:28:18.516925 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:28:19 crc kubenswrapper[4775]: I1126 06:28:19.117912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-vrt6m" Nov 26 06:28:19 crc kubenswrapper[4775]: I1126 06:28:19.138208 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7bwzr" podStartSLOduration=5.449775692 podStartE2EDuration="35.138189267s" podCreationTimestamp="2025-11-26 06:27:44 +0000 UTC" firstStartedPulling="2025-11-26 06:27:45.923822475 +0000 UTC m=+149.285126427" lastFinishedPulling="2025-11-26 06:28:15.61223605 +0000 UTC m=+178.973540002" observedRunningTime="2025-11-26 06:28:16.435995575 +0000 UTC m=+179.797299527" watchObservedRunningTime="2025-11-26 06:28:19.138189267 +0000 UTC m=+182.499493229" Nov 26 06:28:19 crc kubenswrapper[4775]: I1126 06:28:19.356038 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sjc6t" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="registry-server" probeResult="failure" output=< Nov 26 06:28:19 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Nov 26 06:28:19 crc kubenswrapper[4775]: > Nov 26 06:28:19 crc kubenswrapper[4775]: I1126 06:28:19.549275 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rj2bg" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="registry-server" probeResult="failure" output=< Nov 26 06:28:19 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Nov 26 06:28:19 crc kubenswrapper[4775]: > Nov 26 06:28:21 crc kubenswrapper[4775]: I1126 06:28:21.419631 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:28:21 crc kubenswrapper[4775]: I1126 06:28:21.419699 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:28:24 crc kubenswrapper[4775]: I1126 06:28:24.727532 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:28:24 crc kubenswrapper[4775]: I1126 06:28:24.728816 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:28:24 crc kubenswrapper[4775]: I1126 06:28:24.793004 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:28:24 crc kubenswrapper[4775]: I1126 06:28:24.911206 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:28:24 crc kubenswrapper[4775]: I1126 06:28:24.911247 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:28:24 crc kubenswrapper[4775]: I1126 06:28:24.955655 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.110499 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.110543 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.154355 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.338996 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.339031 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.386927 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.403446 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.438237 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.439224 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.441276 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:28:25 crc kubenswrapper[4775]: I1126 06:28:25.449194 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.252549 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cv6pw"] Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.400030 4775 generic.go:334] "Generic (PLEG): container finished" podID="e21090b1-0905-474e-9148-fd349044b039" containerID="718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9" exitCode=0 Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.400077 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hpsd2" event={"ID":"e21090b1-0905-474e-9148-fd349044b039","Type":"ContainerDied","Data":"718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9"} Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.402320 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerStarted","Data":"1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8"} Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.554819 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nk6kf"] Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.555018 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nk6kf" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="registry-server" containerID="cri-o://741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77" gracePeriod=2 Nov 26 06:28:27 crc kubenswrapper[4775]: I1126 06:28:27.916095 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.044760 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-utilities\") pod \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.044825 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88c99\" (UniqueName: \"kubernetes.io/projected/810db382-46fe-4ccc-b05e-5e6f8f540f8e-kube-api-access-88c99\") pod \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.044853 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-catalog-content\") pod \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\" (UID: \"810db382-46fe-4ccc-b05e-5e6f8f540f8e\") " Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.046774 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-utilities" (OuterVolumeSpecName: "utilities") pod "810db382-46fe-4ccc-b05e-5e6f8f540f8e" (UID: "810db382-46fe-4ccc-b05e-5e6f8f540f8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.053451 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/810db382-46fe-4ccc-b05e-5e6f8f540f8e-kube-api-access-88c99" (OuterVolumeSpecName: "kube-api-access-88c99") pod "810db382-46fe-4ccc-b05e-5e6f8f540f8e" (UID: "810db382-46fe-4ccc-b05e-5e6f8f540f8e"). InnerVolumeSpecName "kube-api-access-88c99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.092260 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "810db382-46fe-4ccc-b05e-5e6f8f540f8e" (UID: "810db382-46fe-4ccc-b05e-5e6f8f540f8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.146014 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.146068 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88c99\" (UniqueName: \"kubernetes.io/projected/810db382-46fe-4ccc-b05e-5e6f8f540f8e-kube-api-access-88c99\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.146093 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/810db382-46fe-4ccc-b05e-5e6f8f540f8e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.250556 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.300914 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.409385 4775 generic.go:334] "Generic (PLEG): container finished" podID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerID="741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77" exitCode=0 Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.409458 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6kf" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.409463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6kf" event={"ID":"810db382-46fe-4ccc-b05e-5e6f8f540f8e","Type":"ContainerDied","Data":"741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77"} Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.409597 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6kf" event={"ID":"810db382-46fe-4ccc-b05e-5e6f8f540f8e","Type":"ContainerDied","Data":"0cadb339686dadcd0a2f561a95f5b9c5d5fc5f292c58ee8ad38348f93c6242e8"} Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.409626 4775 scope.go:117] "RemoveContainer" containerID="741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.413127 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hpsd2" event={"ID":"e21090b1-0905-474e-9148-fd349044b039","Type":"ContainerStarted","Data":"58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa"} Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.414730 4775 generic.go:334] "Generic (PLEG): container finished" podID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerID="1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8" exitCode=0 Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.414756 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerDied","Data":"1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8"} Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.432026 4775 scope.go:117] "RemoveContainer" containerID="21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.435000 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hpsd2" podStartSLOduration=2.636797513 podStartE2EDuration="42.434984409s" podCreationTimestamp="2025-11-26 06:27:46 +0000 UTC" firstStartedPulling="2025-11-26 06:27:48.001369625 +0000 UTC m=+151.362673577" lastFinishedPulling="2025-11-26 06:28:27.799556521 +0000 UTC m=+191.160860473" observedRunningTime="2025-11-26 06:28:28.433795627 +0000 UTC m=+191.795099579" watchObservedRunningTime="2025-11-26 06:28:28.434984409 +0000 UTC m=+191.796288361" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.450795 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nk6kf"] Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.453406 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nk6kf"] Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.456814 4775 scope.go:117] "RemoveContainer" containerID="7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.472475 4775 scope.go:117] "RemoveContainer" containerID="741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77" Nov 26 06:28:28 crc kubenswrapper[4775]: E1126 06:28:28.472917 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77\": container with ID starting with 741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77 not found: ID does not exist" containerID="741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.472971 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77"} err="failed to get container status \"741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77\": rpc error: code = NotFound desc = could not find container \"741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77\": container with ID starting with 741c780cd7674178958b32ef90b6ed9ea48df1611421e7f1e362f1bae51d8e77 not found: ID does not exist" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.473023 4775 scope.go:117] "RemoveContainer" containerID="21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1" Nov 26 06:28:28 crc kubenswrapper[4775]: E1126 06:28:28.473330 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1\": container with ID starting with 21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1 not found: ID does not exist" containerID="21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.473359 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1"} err="failed to get container status \"21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1\": rpc error: code = NotFound desc = could not find container \"21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1\": container with ID starting with 21ad8318fd4cd96d7830ebb79638b9cbaf7c3e472ea20258cf0ab7722ad1a3a1 not found: ID does not exist" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.473381 4775 scope.go:117] "RemoveContainer" containerID="7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd" Nov 26 06:28:28 crc kubenswrapper[4775]: E1126 06:28:28.474037 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd\": container with ID starting with 7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd not found: ID does not exist" containerID="7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.474069 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd"} err="failed to get container status \"7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd\": rpc error: code = NotFound desc = could not find container \"7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd\": container with ID starting with 7edaccef6d84acfdd9e572da546116ece96de4c12fd910224b7412782d6104bd not found: ID does not exist" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.550431 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:28:28 crc kubenswrapper[4775]: I1126 06:28:28.592318 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.207841 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 06:28:29 crc kubenswrapper[4775]: E1126 06:28:29.208394 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="extract-utilities" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208416 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="extract-utilities" Nov 26 06:28:29 crc kubenswrapper[4775]: E1126 06:28:29.208430 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95788dae-18a5-49ee-9f1c-8e73527f9673" containerName="pruner" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208437 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="95788dae-18a5-49ee-9f1c-8e73527f9673" containerName="pruner" Nov 26 06:28:29 crc kubenswrapper[4775]: E1126 06:28:29.208455 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="registry-server" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208463 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="registry-server" Nov 26 06:28:29 crc kubenswrapper[4775]: E1126 06:28:29.208473 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="extract-content" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208481 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="extract-content" Nov 26 06:28:29 crc kubenswrapper[4775]: E1126 06:28:29.208499 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36" containerName="pruner" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208507 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36" containerName="pruner" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208618 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" containerName="registry-server" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208635 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="95788dae-18a5-49ee-9f1c-8e73527f9673" containerName="pruner" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.208646 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b2a7d9-a51e-48a3-bbec-3bfb3dc5ae36" containerName="pruner" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.209042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.210431 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.213435 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.215968 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.334162 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="810db382-46fe-4ccc-b05e-5e6f8f540f8e" path="/var/lib/kubelet/pods/810db382-46fe-4ccc-b05e-5e6f8f540f8e/volumes" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.359620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/16839c7c-3d01-4c06-acfc-0678381e02b9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.359674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16839c7c-3d01-4c06-acfc-0678381e02b9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.425804 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerStarted","Data":"2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b"} Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.441646 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5rt8d" podStartSLOduration=2.355770136 podStartE2EDuration="43.441626123s" podCreationTimestamp="2025-11-26 06:27:46 +0000 UTC" firstStartedPulling="2025-11-26 06:27:48.001421176 +0000 UTC m=+151.362725128" lastFinishedPulling="2025-11-26 06:28:29.087277163 +0000 UTC m=+192.448581115" observedRunningTime="2025-11-26 06:28:29.44042279 +0000 UTC m=+192.801726762" watchObservedRunningTime="2025-11-26 06:28:29.441626123 +0000 UTC m=+192.802930095" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.461326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/16839c7c-3d01-4c06-acfc-0678381e02b9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.461420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16839c7c-3d01-4c06-acfc-0678381e02b9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.462608 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/16839c7c-3d01-4c06-acfc-0678381e02b9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.487607 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16839c7c-3d01-4c06-acfc-0678381e02b9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.530289 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.758746 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hx9c6"] Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.759170 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hx9c6" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="registry-server" containerID="cri-o://0645a504dcc3b11fc648460913b693f3c8c441bad0bf3c8ac9417fd09e8b67ee" gracePeriod=2 Nov 26 06:28:29 crc kubenswrapper[4775]: I1126 06:28:29.967964 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.440555 4775 generic.go:334] "Generic (PLEG): container finished" podID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerID="0645a504dcc3b11fc648460913b693f3c8c441bad0bf3c8ac9417fd09e8b67ee" exitCode=0 Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.440617 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx9c6" event={"ID":"69d43e02-199e-48d1-a4a0-a50248a3946e","Type":"ContainerDied","Data":"0645a504dcc3b11fc648460913b693f3c8c441bad0bf3c8ac9417fd09e8b67ee"} Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.442072 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"16839c7c-3d01-4c06-acfc-0678381e02b9","Type":"ContainerStarted","Data":"027f2cd7984d62c06869d1e25bf13d78b8db69d16715d7e66c9c58233cecfb8f"} Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.690837 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.778598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-utilities\") pod \"69d43e02-199e-48d1-a4a0-a50248a3946e\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.778641 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-catalog-content\") pod \"69d43e02-199e-48d1-a4a0-a50248a3946e\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.778695 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bp42\" (UniqueName: \"kubernetes.io/projected/69d43e02-199e-48d1-a4a0-a50248a3946e-kube-api-access-6bp42\") pod \"69d43e02-199e-48d1-a4a0-a50248a3946e\" (UID: \"69d43e02-199e-48d1-a4a0-a50248a3946e\") " Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.783393 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69d43e02-199e-48d1-a4a0-a50248a3946e-kube-api-access-6bp42" (OuterVolumeSpecName: "kube-api-access-6bp42") pod "69d43e02-199e-48d1-a4a0-a50248a3946e" (UID: "69d43e02-199e-48d1-a4a0-a50248a3946e"). InnerVolumeSpecName "kube-api-access-6bp42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.788298 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-utilities" (OuterVolumeSpecName: "utilities") pod "69d43e02-199e-48d1-a4a0-a50248a3946e" (UID: "69d43e02-199e-48d1-a4a0-a50248a3946e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.842562 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69d43e02-199e-48d1-a4a0-a50248a3946e" (UID: "69d43e02-199e-48d1-a4a0-a50248a3946e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.880171 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bp42\" (UniqueName: \"kubernetes.io/projected/69d43e02-199e-48d1-a4a0-a50248a3946e-kube-api-access-6bp42\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.880204 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:30 crc kubenswrapper[4775]: I1126 06:28:30.880216 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69d43e02-199e-48d1-a4a0-a50248a3946e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.448799 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx9c6" event={"ID":"69d43e02-199e-48d1-a4a0-a50248a3946e","Type":"ContainerDied","Data":"46a8d77f522cd4fabf71c903b780b5d47f09c6d1759114493be3bc6406b4a695"} Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.448882 4775 scope.go:117] "RemoveContainer" containerID="0645a504dcc3b11fc648460913b693f3c8c441bad0bf3c8ac9417fd09e8b67ee" Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.448826 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx9c6" Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.450864 4775 generic.go:334] "Generic (PLEG): container finished" podID="16839c7c-3d01-4c06-acfc-0678381e02b9" containerID="9e501377294ccc2571de0b6a4a42269013978261157541ccf291020f919ddcd0" exitCode=0 Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.450898 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"16839c7c-3d01-4c06-acfc-0678381e02b9","Type":"ContainerDied","Data":"9e501377294ccc2571de0b6a4a42269013978261157541ccf291020f919ddcd0"} Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.463248 4775 scope.go:117] "RemoveContainer" containerID="70c16714ab9a989226bcd62795d5eae90c43cd6d86531494ba464912f104be0c" Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.477952 4775 scope.go:117] "RemoveContainer" containerID="2533b128784785dadcd2bee676b6cfa347ab8a2061b2c80f18d1b9a933db53e8" Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.482759 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hx9c6"] Nov 26 06:28:31 crc kubenswrapper[4775]: I1126 06:28:31.482889 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hx9c6"] Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.159016 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rj2bg"] Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.159423 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rj2bg" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="registry-server" containerID="cri-o://a9523fcc45fcd882d1780521daf6d31b4eea97212430d7185c77443d1afb5db1" gracePeriod=2 Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.752439 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.905242 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16839c7c-3d01-4c06-acfc-0678381e02b9-kube-api-access\") pod \"16839c7c-3d01-4c06-acfc-0678381e02b9\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.905346 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/16839c7c-3d01-4c06-acfc-0678381e02b9-kubelet-dir\") pod \"16839c7c-3d01-4c06-acfc-0678381e02b9\" (UID: \"16839c7c-3d01-4c06-acfc-0678381e02b9\") " Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.905549 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/16839c7c-3d01-4c06-acfc-0678381e02b9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "16839c7c-3d01-4c06-acfc-0678381e02b9" (UID: "16839c7c-3d01-4c06-acfc-0678381e02b9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:28:32 crc kubenswrapper[4775]: I1126 06:28:32.914900 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16839c7c-3d01-4c06-acfc-0678381e02b9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "16839c7c-3d01-4c06-acfc-0678381e02b9" (UID: "16839c7c-3d01-4c06-acfc-0678381e02b9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.006317 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16839c7c-3d01-4c06-acfc-0678381e02b9-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.006352 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/16839c7c-3d01-4c06-acfc-0678381e02b9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.333741 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" path="/var/lib/kubelet/pods/69d43e02-199e-48d1-a4a0-a50248a3946e/volumes" Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.475188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"16839c7c-3d01-4c06-acfc-0678381e02b9","Type":"ContainerDied","Data":"027f2cd7984d62c06869d1e25bf13d78b8db69d16715d7e66c9c58233cecfb8f"} Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.475626 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="027f2cd7984d62c06869d1e25bf13d78b8db69d16715d7e66c9c58233cecfb8f" Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.475221 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.477898 4775 generic.go:334] "Generic (PLEG): container finished" podID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerID="a9523fcc45fcd882d1780521daf6d31b4eea97212430d7185c77443d1afb5db1" exitCode=0 Nov 26 06:28:33 crc kubenswrapper[4775]: I1126 06:28:33.477960 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerDied","Data":"a9523fcc45fcd882d1780521daf6d31b4eea97212430d7185c77443d1afb5db1"} Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203174 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 06:28:34 crc kubenswrapper[4775]: E1126 06:28:34.203415 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="registry-server" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203430 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="registry-server" Nov 26 06:28:34 crc kubenswrapper[4775]: E1126 06:28:34.203444 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="extract-utilities" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203452 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="extract-utilities" Nov 26 06:28:34 crc kubenswrapper[4775]: E1126 06:28:34.203476 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16839c7c-3d01-4c06-acfc-0678381e02b9" containerName="pruner" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203485 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="16839c7c-3d01-4c06-acfc-0678381e02b9" containerName="pruner" Nov 26 06:28:34 crc kubenswrapper[4775]: E1126 06:28:34.203500 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="extract-content" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203509 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="extract-content" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203622 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="16839c7c-3d01-4c06-acfc-0678381e02b9" containerName="pruner" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.203636 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="69d43e02-199e-48d1-a4a0-a50248a3946e" containerName="registry-server" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.204145 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.208260 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.208678 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.233503 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.331267 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-var-lock\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.331349 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.331866 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228b00f8-984b-43be-a5ec-99e8414ee44c-kube-api-access\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.432428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-var-lock\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.432779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.432821 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228b00f8-984b-43be-a5ec-99e8414ee44c-kube-api-access\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.432870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-kubelet-dir\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.432568 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-var-lock\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.452224 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228b00f8-984b-43be-a5ec-99e8414ee44c-kube-api-access\") pod \"installer-9-crc\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.484695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rj2bg" event={"ID":"f5dc821f-58e0-47ec-8aac-d67b333013a4","Type":"ContainerDied","Data":"6bedf77141fde9cb742b821814f34891a658dbfdffd398378d525a0546464261"} Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.484748 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bedf77141fde9cb742b821814f34891a658dbfdffd398378d525a0546464261" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.492669 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.555116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.634573 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-utilities\") pod \"f5dc821f-58e0-47ec-8aac-d67b333013a4\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.634807 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-catalog-content\") pod \"f5dc821f-58e0-47ec-8aac-d67b333013a4\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.634909 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq9wt\" (UniqueName: \"kubernetes.io/projected/f5dc821f-58e0-47ec-8aac-d67b333013a4-kube-api-access-wq9wt\") pod \"f5dc821f-58e0-47ec-8aac-d67b333013a4\" (UID: \"f5dc821f-58e0-47ec-8aac-d67b333013a4\") " Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.635245 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-utilities" (OuterVolumeSpecName: "utilities") pod "f5dc821f-58e0-47ec-8aac-d67b333013a4" (UID: "f5dc821f-58e0-47ec-8aac-d67b333013a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.635374 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.654198 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5dc821f-58e0-47ec-8aac-d67b333013a4-kube-api-access-wq9wt" (OuterVolumeSpecName: "kube-api-access-wq9wt") pod "f5dc821f-58e0-47ec-8aac-d67b333013a4" (UID: "f5dc821f-58e0-47ec-8aac-d67b333013a4"). InnerVolumeSpecName "kube-api-access-wq9wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.736128 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq9wt\" (UniqueName: \"kubernetes.io/projected/f5dc821f-58e0-47ec-8aac-d67b333013a4-kube-api-access-wq9wt\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.743095 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5dc821f-58e0-47ec-8aac-d67b333013a4" (UID: "f5dc821f-58e0-47ec-8aac-d67b333013a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.836824 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5dc821f-58e0-47ec-8aac-d67b333013a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:34 crc kubenswrapper[4775]: I1126 06:28:34.970545 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 06:28:34 crc kubenswrapper[4775]: W1126 06:28:34.982956 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod228b00f8_984b_43be_a5ec_99e8414ee44c.slice/crio-a4b17b6d6e119f4339b5d5c821cf11ea11bc25a107526fae1997e53d952c5643 WatchSource:0}: Error finding container a4b17b6d6e119f4339b5d5c821cf11ea11bc25a107526fae1997e53d952c5643: Status 404 returned error can't find the container with id a4b17b6d6e119f4339b5d5c821cf11ea11bc25a107526fae1997e53d952c5643 Nov 26 06:28:35 crc kubenswrapper[4775]: I1126 06:28:35.491797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"228b00f8-984b-43be-a5ec-99e8414ee44c","Type":"ContainerStarted","Data":"a4b17b6d6e119f4339b5d5c821cf11ea11bc25a107526fae1997e53d952c5643"} Nov 26 06:28:35 crc kubenswrapper[4775]: I1126 06:28:35.491841 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rj2bg" Nov 26 06:28:35 crc kubenswrapper[4775]: I1126 06:28:35.516831 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rj2bg"] Nov 26 06:28:35 crc kubenswrapper[4775]: I1126 06:28:35.521930 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rj2bg"] Nov 26 06:28:36 crc kubenswrapper[4775]: I1126 06:28:36.497795 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"228b00f8-984b-43be-a5ec-99e8414ee44c","Type":"ContainerStarted","Data":"a1562b1d5d21fe36ca9d15f331fc392945ca57e6727138a578799eb092cfc8a8"} Nov 26 06:28:36 crc kubenswrapper[4775]: I1126 06:28:36.517702 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.517685432 podStartE2EDuration="2.517685432s" podCreationTimestamp="2025-11-26 06:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:28:36.515311096 +0000 UTC m=+199.876615068" watchObservedRunningTime="2025-11-26 06:28:36.517685432 +0000 UTC m=+199.878989384" Nov 26 06:28:36 crc kubenswrapper[4775]: I1126 06:28:36.754372 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:28:36 crc kubenswrapper[4775]: I1126 06:28:36.754419 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:28:36 crc kubenswrapper[4775]: I1126 06:28:36.794504 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:28:37 crc kubenswrapper[4775]: I1126 06:28:37.146778 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:28:37 crc kubenswrapper[4775]: I1126 06:28:37.147257 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:28:37 crc kubenswrapper[4775]: I1126 06:28:37.218369 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:28:37 crc kubenswrapper[4775]: I1126 06:28:37.338728 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" path="/var/lib/kubelet/pods/f5dc821f-58e0-47ec-8aac-d67b333013a4/volumes" Nov 26 06:28:37 crc kubenswrapper[4775]: I1126 06:28:37.552345 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:28:37 crc kubenswrapper[4775]: I1126 06:28:37.575385 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:28:40 crc kubenswrapper[4775]: I1126 06:28:40.562074 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rt8d"] Nov 26 06:28:40 crc kubenswrapper[4775]: I1126 06:28:40.564269 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5rt8d" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="registry-server" containerID="cri-o://2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b" gracePeriod=2 Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.035138 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.043116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-utilities\") pod \"eaf163ed-6550-4efb-b83e-8fd80bea0935\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.043151 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-catalog-content\") pod \"eaf163ed-6550-4efb-b83e-8fd80bea0935\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.043186 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95njv\" (UniqueName: \"kubernetes.io/projected/eaf163ed-6550-4efb-b83e-8fd80bea0935-kube-api-access-95njv\") pod \"eaf163ed-6550-4efb-b83e-8fd80bea0935\" (UID: \"eaf163ed-6550-4efb-b83e-8fd80bea0935\") " Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.044782 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-utilities" (OuterVolumeSpecName: "utilities") pod "eaf163ed-6550-4efb-b83e-8fd80bea0935" (UID: "eaf163ed-6550-4efb-b83e-8fd80bea0935"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.053962 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf163ed-6550-4efb-b83e-8fd80bea0935-kube-api-access-95njv" (OuterVolumeSpecName: "kube-api-access-95njv") pod "eaf163ed-6550-4efb-b83e-8fd80bea0935" (UID: "eaf163ed-6550-4efb-b83e-8fd80bea0935"). InnerVolumeSpecName "kube-api-access-95njv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.066389 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eaf163ed-6550-4efb-b83e-8fd80bea0935" (UID: "eaf163ed-6550-4efb-b83e-8fd80bea0935"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.144825 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95njv\" (UniqueName: \"kubernetes.io/projected/eaf163ed-6550-4efb-b83e-8fd80bea0935-kube-api-access-95njv\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.144858 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.144869 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaf163ed-6550-4efb-b83e-8fd80bea0935-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.542966 4775 generic.go:334] "Generic (PLEG): container finished" podID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerID="2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b" exitCode=0 Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.543046 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rt8d" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.543074 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerDied","Data":"2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b"} Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.543555 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rt8d" event={"ID":"eaf163ed-6550-4efb-b83e-8fd80bea0935","Type":"ContainerDied","Data":"296f753163ada7f2f4e6130686d1898889068ce43c17eae8d23216fdaadb7635"} Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.543606 4775 scope.go:117] "RemoveContainer" containerID="2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.583829 4775 scope.go:117] "RemoveContainer" containerID="1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.585154 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rt8d"] Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.592060 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rt8d"] Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.614205 4775 scope.go:117] "RemoveContainer" containerID="f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.646870 4775 scope.go:117] "RemoveContainer" containerID="2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b" Nov 26 06:28:41 crc kubenswrapper[4775]: E1126 06:28:41.647522 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b\": container with ID starting with 2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b not found: ID does not exist" containerID="2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.647567 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b"} err="failed to get container status \"2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b\": rpc error: code = NotFound desc = could not find container \"2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b\": container with ID starting with 2e584322f44ab43f39b3209643f89c3eee8b6cae1149fb90448434516912a72b not found: ID does not exist" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.647600 4775 scope.go:117] "RemoveContainer" containerID="1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8" Nov 26 06:28:41 crc kubenswrapper[4775]: E1126 06:28:41.648031 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8\": container with ID starting with 1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8 not found: ID does not exist" containerID="1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.648103 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8"} err="failed to get container status \"1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8\": rpc error: code = NotFound desc = could not find container \"1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8\": container with ID starting with 1129b04b1f8d7c5cd9954b8b362eb3806cfcf1c1fc2cbc220c24e6b489e511d8 not found: ID does not exist" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.648146 4775 scope.go:117] "RemoveContainer" containerID="f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255" Nov 26 06:28:41 crc kubenswrapper[4775]: E1126 06:28:41.648645 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255\": container with ID starting with f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255 not found: ID does not exist" containerID="f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255" Nov 26 06:28:41 crc kubenswrapper[4775]: I1126 06:28:41.648683 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255"} err="failed to get container status \"f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255\": rpc error: code = NotFound desc = could not find container \"f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255\": container with ID starting with f7be0dbdab5a7253981f7fe12078b65875636233cb24c7418299f560ed644255 not found: ID does not exist" Nov 26 06:28:43 crc kubenswrapper[4775]: I1126 06:28:43.334650 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" path="/var/lib/kubelet/pods/eaf163ed-6550-4efb-b83e-8fd80bea0935/volumes" Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.419697 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.420354 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.420428 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.421313 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.421421 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f" gracePeriod=600 Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.607213 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f" exitCode=0 Nov 26 06:28:51 crc kubenswrapper[4775]: I1126 06:28:51.607321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f"} Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.289470 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" podUID="61d701be-8e8d-4d60-a562-328010fc2003" containerName="oauth-openshift" containerID="cri-o://31822ae474dcc4eee059f2f0422b6306e243f86f5f35c6c6113d280ff03c89d1" gracePeriod=15 Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.615006 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"a22b92d95ea00f90fd226634f524ff08e52cb46e50050a0f9a3125f6b85b48ba"} Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.619333 4775 generic.go:334] "Generic (PLEG): container finished" podID="61d701be-8e8d-4d60-a562-328010fc2003" containerID="31822ae474dcc4eee059f2f0422b6306e243f86f5f35c6c6113d280ff03c89d1" exitCode=0 Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.619395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" event={"ID":"61d701be-8e8d-4d60-a562-328010fc2003","Type":"ContainerDied","Data":"31822ae474dcc4eee059f2f0422b6306e243f86f5f35c6c6113d280ff03c89d1"} Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.670909 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.801874 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-audit-policies\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.801958 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-trusted-ca-bundle\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802024 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-idp-0-file-data\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802075 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-provider-selection\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802120 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-ocp-branding-template\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802153 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtdqd\" (UniqueName: \"kubernetes.io/projected/61d701be-8e8d-4d60-a562-328010fc2003-kube-api-access-xtdqd\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802198 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-login\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802233 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-session\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802271 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-router-certs\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802320 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-cliconfig\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802369 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-error\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802413 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-service-ca\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802444 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-serving-cert\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.802478 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61d701be-8e8d-4d60-a562-328010fc2003-audit-dir\") pod \"61d701be-8e8d-4d60-a562-328010fc2003\" (UID: \"61d701be-8e8d-4d60-a562-328010fc2003\") " Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.803004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.803180 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.804245 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.804492 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61d701be-8e8d-4d60-a562-328010fc2003-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.806774 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.810429 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.810917 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.811265 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.811522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d701be-8e8d-4d60-a562-328010fc2003-kube-api-access-xtdqd" (OuterVolumeSpecName: "kube-api-access-xtdqd") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "kube-api-access-xtdqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.811773 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.812021 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.812309 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.812920 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.816015 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "61d701be-8e8d-4d60-a562-328010fc2003" (UID: "61d701be-8e8d-4d60-a562-328010fc2003"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904356 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904394 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtdqd\" (UniqueName: \"kubernetes.io/projected/61d701be-8e8d-4d60-a562-328010fc2003-kube-api-access-xtdqd\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904405 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904416 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904424 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904432 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904441 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904450 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904481 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904491 4775 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61d701be-8e8d-4d60-a562-328010fc2003-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904500 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904510 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904518 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:52 crc kubenswrapper[4775]: I1126 06:28:52.904527 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/61d701be-8e8d-4d60-a562-328010fc2003-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.417257 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5494594499-z4wtq"] Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.417998 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="extract-content" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418030 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="extract-content" Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.418057 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d701be-8e8d-4d60-a562-328010fc2003" containerName="oauth-openshift" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418072 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d701be-8e8d-4d60-a562-328010fc2003" containerName="oauth-openshift" Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.418091 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="registry-server" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418104 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="registry-server" Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.418127 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="registry-server" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418139 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="registry-server" Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.418154 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="extract-utilities" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418188 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="extract-utilities" Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.418209 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="extract-content" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418221 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="extract-content" Nov 26 06:28:53 crc kubenswrapper[4775]: E1126 06:28:53.418233 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="extract-utilities" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418246 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="extract-utilities" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418428 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5dc821f-58e0-47ec-8aac-d67b333013a4" containerName="registry-server" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418464 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d701be-8e8d-4d60-a562-328010fc2003" containerName="oauth-openshift" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.418491 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf163ed-6550-4efb-b83e-8fd80bea0935" containerName="registry-server" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.419310 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.431619 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5494594499-z4wtq"] Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-session\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614332 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614358 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-audit-policies\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614395 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614936 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-service-ca\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.614994 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615051 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7mrr\" (UniqueName: \"kubernetes.io/projected/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-kube-api-access-h7mrr\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-audit-dir\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-login\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-error\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615397 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615431 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-router-certs\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.615496 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.628449 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.628436 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cv6pw" event={"ID":"61d701be-8e8d-4d60-a562-328010fc2003","Type":"ContainerDied","Data":"a5e22261a7db7e32d5607a55a8fd75f95373c95c0b2d671f90456e415682bc20"} Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.628545 4775 scope.go:117] "RemoveContainer" containerID="31822ae474dcc4eee059f2f0422b6306e243f86f5f35c6c6113d280ff03c89d1" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.655293 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cv6pw"] Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.660941 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cv6pw"] Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.716486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.716543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-audit-policies\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.716571 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.716598 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-service-ca\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.716966 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7mrr\" (UniqueName: \"kubernetes.io/projected/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-kube-api-access-h7mrr\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-audit-dir\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-login\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717474 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-error\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717537 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717574 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-audit-dir\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-router-certs\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717755 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-session\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.717871 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.720697 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-audit-policies\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.721445 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.723064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.724055 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-service-ca\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.724176 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-login\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.724798 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-error\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.725128 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-session\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.726492 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.727490 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.729072 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.729521 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.732401 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-v4-0-config-system-router-certs\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.744161 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7mrr\" (UniqueName: \"kubernetes.io/projected/a138c9a7-37c3-4dc5-a3db-9e4d9a048c32-kube-api-access-h7mrr\") pod \"oauth-openshift-5494594499-z4wtq\" (UID: \"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32\") " pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.761992 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:53 crc kubenswrapper[4775]: I1126 06:28:53.977776 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5494594499-z4wtq"] Nov 26 06:28:53 crc kubenswrapper[4775]: W1126 06:28:53.983771 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda138c9a7_37c3_4dc5_a3db_9e4d9a048c32.slice/crio-d2a5741a90cb884bbfe8347ad5cb5f195d517709cd7693881cdf1b1e71aba260 WatchSource:0}: Error finding container d2a5741a90cb884bbfe8347ad5cb5f195d517709cd7693881cdf1b1e71aba260: Status 404 returned error can't find the container with id d2a5741a90cb884bbfe8347ad5cb5f195d517709cd7693881cdf1b1e71aba260 Nov 26 06:28:54 crc kubenswrapper[4775]: I1126 06:28:54.635231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" event={"ID":"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32","Type":"ContainerStarted","Data":"fd61a9687f0ec568d86cc14533d6b2913c67ed455bdbe5fed2d209668aebeccb"} Nov 26 06:28:54 crc kubenswrapper[4775]: I1126 06:28:54.636620 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:54 crc kubenswrapper[4775]: I1126 06:28:54.636653 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" event={"ID":"a138c9a7-37c3-4dc5-a3db-9e4d9a048c32","Type":"ContainerStarted","Data":"d2a5741a90cb884bbfe8347ad5cb5f195d517709cd7693881cdf1b1e71aba260"} Nov 26 06:28:54 crc kubenswrapper[4775]: I1126 06:28:54.917588 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" Nov 26 06:28:54 crc kubenswrapper[4775]: I1126 06:28:54.945290 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5494594499-z4wtq" podStartSLOduration=27.945273515 podStartE2EDuration="27.945273515s" podCreationTimestamp="2025-11-26 06:28:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:28:54.653010049 +0000 UTC m=+218.014314031" watchObservedRunningTime="2025-11-26 06:28:54.945273515 +0000 UTC m=+218.306577477" Nov 26 06:28:55 crc kubenswrapper[4775]: I1126 06:28:55.334193 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d701be-8e8d-4d60-a562-328010fc2003" path="/var/lib/kubelet/pods/61d701be-8e8d-4d60-a562-328010fc2003/volumes" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.085818 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7bwzr"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.086464 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7bwzr" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="registry-server" containerID="cri-o://aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e" gracePeriod=30 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.095815 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9p8qj"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.096105 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9p8qj" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="registry-server" containerID="cri-o://34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e" gracePeriod=30 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.102117 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z59ck"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.102356 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerName="marketplace-operator" containerID="cri-o://90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3" gracePeriod=30 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.115559 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hpsd2"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.115808 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hpsd2" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="registry-server" containerID="cri-o://58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa" gracePeriod=30 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.119685 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sjc6t"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.120136 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sjc6t" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="registry-server" containerID="cri-o://a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36" gracePeriod=30 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.121727 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sh8wn"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.122332 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.132666 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sh8wn"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.289815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c5e343-455f-43ad-9936-8b64224702bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.289881 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4c5e343-455f-43ad-9936-8b64224702bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.289906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65tvl\" (UniqueName: \"kubernetes.io/projected/a4c5e343-455f-43ad-9936-8b64224702bf-kube-api-access-65tvl\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.391251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c5e343-455f-43ad-9936-8b64224702bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.391488 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4c5e343-455f-43ad-9936-8b64224702bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.391506 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65tvl\" (UniqueName: \"kubernetes.io/projected/a4c5e343-455f-43ad-9936-8b64224702bf-kube-api-access-65tvl\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.396276 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c5e343-455f-43ad-9936-8b64224702bf-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.409226 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4c5e343-455f-43ad-9936-8b64224702bf-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.409244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65tvl\" (UniqueName: \"kubernetes.io/projected/a4c5e343-455f-43ad-9936-8b64224702bf-kube-api-access-65tvl\") pod \"marketplace-operator-79b997595-sh8wn\" (UID: \"a4c5e343-455f-43ad-9936-8b64224702bf\") " pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.451549 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.551094 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.570621 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.601003 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-catalog-content\") pod \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.601061 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-utilities\") pod \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.601203 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-catalog-content\") pod \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.601309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qfb6\" (UniqueName: \"kubernetes.io/projected/4bdf9960-b1db-4086-9492-1b656cbe6e1b-kube-api-access-4qfb6\") pod \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.601362 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fbbn\" (UniqueName: \"kubernetes.io/projected/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-kube-api-access-8fbbn\") pod \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\" (UID: \"7ea68cee-a3d6-4ff7-b668-d87599fa4efd\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.601414 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-utilities\") pod \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\" (UID: \"4bdf9960-b1db-4086-9492-1b656cbe6e1b\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.604887 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bdf9960-b1db-4086-9492-1b656cbe6e1b-kube-api-access-4qfb6" (OuterVolumeSpecName: "kube-api-access-4qfb6") pod "4bdf9960-b1db-4086-9492-1b656cbe6e1b" (UID: "4bdf9960-b1db-4086-9492-1b656cbe6e1b"). InnerVolumeSpecName "kube-api-access-4qfb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.606632 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-utilities" (OuterVolumeSpecName: "utilities") pod "7ea68cee-a3d6-4ff7-b668-d87599fa4efd" (UID: "7ea68cee-a3d6-4ff7-b668-d87599fa4efd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.607440 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-utilities" (OuterVolumeSpecName: "utilities") pod "4bdf9960-b1db-4086-9492-1b656cbe6e1b" (UID: "4bdf9960-b1db-4086-9492-1b656cbe6e1b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.612065 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.612335 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.612510 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qfb6\" (UniqueName: \"kubernetes.io/projected/4bdf9960-b1db-4086-9492-1b656cbe6e1b-kube-api-access-4qfb6\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.612451 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-kube-api-access-8fbbn" (OuterVolumeSpecName: "kube-api-access-8fbbn") pod "7ea68cee-a3d6-4ff7-b668-d87599fa4efd" (UID: "7ea68cee-a3d6-4ff7-b668-d87599fa4efd"). InnerVolumeSpecName "kube-api-access-8fbbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.629354 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.641383 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.641653 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.711012 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ea68cee-a3d6-4ff7-b668-d87599fa4efd" (UID: "7ea68cee-a3d6-4ff7-b668-d87599fa4efd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.711947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bdf9960-b1db-4086-9492-1b656cbe6e1b" (UID: "4bdf9960-b1db-4086-9492-1b656cbe6e1b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713408 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kxrj\" (UniqueName: \"kubernetes.io/projected/e21090b1-0905-474e-9148-fd349044b039-kube-api-access-7kxrj\") pod \"e21090b1-0905-474e-9148-fd349044b039\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-utilities\") pod \"64dc0036-294e-4f2b-9b9d-cd540c42e558\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713480 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-utilities\") pod \"e21090b1-0905-474e-9148-fd349044b039\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713559 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl5n7\" (UniqueName: \"kubernetes.io/projected/f1365993-a452-4e00-92e0-64e1c08ce1c2-kube-api-access-bl5n7\") pod \"f1365993-a452-4e00-92e0-64e1c08ce1c2\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713588 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zc2z\" (UniqueName: \"kubernetes.io/projected/64dc0036-294e-4f2b-9b9d-cd540c42e558-kube-api-access-8zc2z\") pod \"64dc0036-294e-4f2b-9b9d-cd540c42e558\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713605 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-catalog-content\") pod \"64dc0036-294e-4f2b-9b9d-cd540c42e558\" (UID: \"64dc0036-294e-4f2b-9b9d-cd540c42e558\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713628 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-operator-metrics\") pod \"f1365993-a452-4e00-92e0-64e1c08ce1c2\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713648 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-catalog-content\") pod \"e21090b1-0905-474e-9148-fd349044b039\" (UID: \"e21090b1-0905-474e-9148-fd349044b039\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713678 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-trusted-ca\") pod \"f1365993-a452-4e00-92e0-64e1c08ce1c2\" (UID: \"f1365993-a452-4e00-92e0-64e1c08ce1c2\") " Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713901 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdf9960-b1db-4086-9492-1b656cbe6e1b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713917 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.713951 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fbbn\" (UniqueName: \"kubernetes.io/projected/7ea68cee-a3d6-4ff7-b668-d87599fa4efd-kube-api-access-8fbbn\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.714494 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f1365993-a452-4e00-92e0-64e1c08ce1c2" (UID: "f1365993-a452-4e00-92e0-64e1c08ce1c2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.717569 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e21090b1-0905-474e-9148-fd349044b039-kube-api-access-7kxrj" (OuterVolumeSpecName: "kube-api-access-7kxrj") pod "e21090b1-0905-474e-9148-fd349044b039" (UID: "e21090b1-0905-474e-9148-fd349044b039"). InnerVolumeSpecName "kube-api-access-7kxrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.721338 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f1365993-a452-4e00-92e0-64e1c08ce1c2" (UID: "f1365993-a452-4e00-92e0-64e1c08ce1c2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.722344 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-utilities" (OuterVolumeSpecName: "utilities") pod "e21090b1-0905-474e-9148-fd349044b039" (UID: "e21090b1-0905-474e-9148-fd349044b039"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.722970 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64dc0036-294e-4f2b-9b9d-cd540c42e558-kube-api-access-8zc2z" (OuterVolumeSpecName: "kube-api-access-8zc2z") pod "64dc0036-294e-4f2b-9b9d-cd540c42e558" (UID: "64dc0036-294e-4f2b-9b9d-cd540c42e558"). InnerVolumeSpecName "kube-api-access-8zc2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.723920 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1365993-a452-4e00-92e0-64e1c08ce1c2-kube-api-access-bl5n7" (OuterVolumeSpecName: "kube-api-access-bl5n7") pod "f1365993-a452-4e00-92e0-64e1c08ce1c2" (UID: "f1365993-a452-4e00-92e0-64e1c08ce1c2"). InnerVolumeSpecName "kube-api-access-bl5n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.725879 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-utilities" (OuterVolumeSpecName: "utilities") pod "64dc0036-294e-4f2b-9b9d-cd540c42e558" (UID: "64dc0036-294e-4f2b-9b9d-cd540c42e558"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.728950 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerID="aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e" exitCode=0 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.729023 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bwzr" event={"ID":"7ea68cee-a3d6-4ff7-b668-d87599fa4efd","Type":"ContainerDied","Data":"aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.729049 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7bwzr" event={"ID":"7ea68cee-a3d6-4ff7-b668-d87599fa4efd","Type":"ContainerDied","Data":"6134b8cb59ee28ccc393b5fa15e7597ec74d6a6a50c2eebbd117bd4de9fac7b1"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.729066 4775 scope.go:117] "RemoveContainer" containerID="aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.729169 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7bwzr" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.732622 4775 generic.go:334] "Generic (PLEG): container finished" podID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerID="90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3" exitCode=0 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.732695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" event={"ID":"f1365993-a452-4e00-92e0-64e1c08ce1c2","Type":"ContainerDied","Data":"90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.732744 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" event={"ID":"f1365993-a452-4e00-92e0-64e1c08ce1c2","Type":"ContainerDied","Data":"4c8b9705a97ef316c2823683118950056b723dbb9cf9bf9bad5e7378de964bee"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.732812 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z59ck" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.737826 4775 generic.go:334] "Generic (PLEG): container finished" podID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerID="a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36" exitCode=0 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.738160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerDied","Data":"a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.738193 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sjc6t" event={"ID":"64dc0036-294e-4f2b-9b9d-cd540c42e558","Type":"ContainerDied","Data":"b81c3b3f3fcbcb6c5e891e4c9ef3e469ae6b0344e5908e0041128b63f1295ddf"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.738271 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sjc6t" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.742252 4775 generic.go:334] "Generic (PLEG): container finished" podID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerID="34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e" exitCode=0 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.742374 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p8qj" event={"ID":"4bdf9960-b1db-4086-9492-1b656cbe6e1b","Type":"ContainerDied","Data":"34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.742408 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p8qj" event={"ID":"4bdf9960-b1db-4086-9492-1b656cbe6e1b","Type":"ContainerDied","Data":"1545686d3e732e381cd24ae74bd3fd250837d3d115bb5afce042439223d7b43e"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.742536 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p8qj" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.750619 4775 generic.go:334] "Generic (PLEG): container finished" podID="e21090b1-0905-474e-9148-fd349044b039" containerID="58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa" exitCode=0 Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.750670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hpsd2" event={"ID":"e21090b1-0905-474e-9148-fd349044b039","Type":"ContainerDied","Data":"58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.750678 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hpsd2" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.750720 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hpsd2" event={"ID":"e21090b1-0905-474e-9148-fd349044b039","Type":"ContainerDied","Data":"1767fff75c505352c7aa2f8cdff6b0f5fef44e2000ced6fa49aa7fa5a14e9e0a"} Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.757670 4775 scope.go:117] "RemoveContainer" containerID="18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.807471 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z59ck"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.808393 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e21090b1-0905-474e-9148-fd349044b039" (UID: "e21090b1-0905-474e-9148-fd349044b039"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.813402 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z59ck"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.814440 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7bwzr"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.818024 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7bwzr"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.819978 4775 scope.go:117] "RemoveContainer" containerID="64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820842 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl5n7\" (UniqueName: \"kubernetes.io/projected/f1365993-a452-4e00-92e0-64e1c08ce1c2-kube-api-access-bl5n7\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820870 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zc2z\" (UniqueName: \"kubernetes.io/projected/64dc0036-294e-4f2b-9b9d-cd540c42e558-kube-api-access-8zc2z\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820885 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820899 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820911 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f1365993-a452-4e00-92e0-64e1c08ce1c2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820923 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kxrj\" (UniqueName: \"kubernetes.io/projected/e21090b1-0905-474e-9148-fd349044b039-kube-api-access-7kxrj\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820934 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.820946 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e21090b1-0905-474e-9148-fd349044b039-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.828737 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9p8qj"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.830177 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9p8qj"] Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.858934 4775 scope.go:117] "RemoveContainer" containerID="aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.863999 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e\": container with ID starting with aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e not found: ID does not exist" containerID="aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.864045 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e"} err="failed to get container status \"aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e\": rpc error: code = NotFound desc = could not find container \"aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e\": container with ID starting with aefdb10c4522f645342bfcb770901bd7c8139101bafd9d7e9a7ff2d6e6e63a9e not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.864075 4775 scope.go:117] "RemoveContainer" containerID="18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.864419 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a\": container with ID starting with 18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a not found: ID does not exist" containerID="18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.864444 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a"} err="failed to get container status \"18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a\": rpc error: code = NotFound desc = could not find container \"18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a\": container with ID starting with 18dfb2db5e0cf3a6a90dc7d9cef64b9ee91e166a756841166c6c1c319b6da69a not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.864459 4775 scope.go:117] "RemoveContainer" containerID="64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.864682 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd\": container with ID starting with 64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd not found: ID does not exist" containerID="64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.864703 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd"} err="failed to get container status \"64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd\": rpc error: code = NotFound desc = could not find container \"64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd\": container with ID starting with 64c41f3eda305c803fff0c638b55093d58280364c8f358923efa3910b534c6cd not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.864768 4775 scope.go:117] "RemoveContainer" containerID="90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.889602 4775 scope.go:117] "RemoveContainer" containerID="90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.890081 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3\": container with ID starting with 90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3 not found: ID does not exist" containerID="90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.890112 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3"} err="failed to get container status \"90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3\": rpc error: code = NotFound desc = could not find container \"90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3\": container with ID starting with 90eed34c4ca67baee142c660da7f62d6b57c4088ce8aeeba23a0c7016c45c7d3 not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.890133 4775 scope.go:117] "RemoveContainer" containerID="a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.902966 4775 scope.go:117] "RemoveContainer" containerID="984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.922131 4775 scope.go:117] "RemoveContainer" containerID="de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.935409 4775 scope.go:117] "RemoveContainer" containerID="a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.935782 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36\": container with ID starting with a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36 not found: ID does not exist" containerID="a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.935982 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36"} err="failed to get container status \"a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36\": rpc error: code = NotFound desc = could not find container \"a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36\": container with ID starting with a4fc6deabdfd717f5394065bf3ac42d3e508c808edf1558f81a2bb0aaccadc36 not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.936019 4775 scope.go:117] "RemoveContainer" containerID="984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.936373 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa\": container with ID starting with 984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa not found: ID does not exist" containerID="984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.936403 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa"} err="failed to get container status \"984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa\": rpc error: code = NotFound desc = could not find container \"984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa\": container with ID starting with 984ac297774399777888f416146004061f94155b765d67f79fc1919c6fb129fa not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.936422 4775 scope.go:117] "RemoveContainer" containerID="de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.936683 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1\": container with ID starting with de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1 not found: ID does not exist" containerID="de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.936699 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1"} err="failed to get container status \"de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1\": rpc error: code = NotFound desc = could not find container \"de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1\": container with ID starting with de240852acc85e2904cd6893a1818d266a39fec0ff24c0983dad1a0d23f783f1 not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.936729 4775 scope.go:117] "RemoveContainer" containerID="34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.940248 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64dc0036-294e-4f2b-9b9d-cd540c42e558" (UID: "64dc0036-294e-4f2b-9b9d-cd540c42e558"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.948394 4775 scope.go:117] "RemoveContainer" containerID="176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.960867 4775 scope.go:117] "RemoveContainer" containerID="446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.973853 4775 scope.go:117] "RemoveContainer" containerID="34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.974318 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e\": container with ID starting with 34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e not found: ID does not exist" containerID="34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.974356 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e"} err="failed to get container status \"34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e\": rpc error: code = NotFound desc = could not find container \"34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e\": container with ID starting with 34a64be660668114299fa04b6c9270f81899bad6fb0dadd26093f18c0e92844e not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.974385 4775 scope.go:117] "RemoveContainer" containerID="176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.974656 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f\": container with ID starting with 176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f not found: ID does not exist" containerID="176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.974885 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f"} err="failed to get container status \"176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f\": rpc error: code = NotFound desc = could not find container \"176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f\": container with ID starting with 176f554ccd53ccc2843b7b7449de898f96e6595063ead2aa09241a2bb6b4921f not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.974914 4775 scope.go:117] "RemoveContainer" containerID="446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59" Nov 26 06:29:07 crc kubenswrapper[4775]: E1126 06:29:07.982529 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59\": container with ID starting with 446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59 not found: ID does not exist" containerID="446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.982573 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59"} err="failed to get container status \"446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59\": rpc error: code = NotFound desc = could not find container \"446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59\": container with ID starting with 446dd25808c8bbd6f4d9aa4f8e194a03a9babd0edeea2207a1ce6deb34d77b59 not found: ID does not exist" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.982598 4775 scope.go:117] "RemoveContainer" containerID="58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa" Nov 26 06:29:07 crc kubenswrapper[4775]: I1126 06:29:07.995211 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sh8wn"] Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.000771 4775 scope.go:117] "RemoveContainer" containerID="718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9" Nov 26 06:29:08 crc kubenswrapper[4775]: W1126 06:29:08.004342 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4c5e343_455f_43ad_9936_8b64224702bf.slice/crio-655247bd6140e4a06335ad8537904d593fe054bdd9527636886c20e42cd100c1 WatchSource:0}: Error finding container 655247bd6140e4a06335ad8537904d593fe054bdd9527636886c20e42cd100c1: Status 404 returned error can't find the container with id 655247bd6140e4a06335ad8537904d593fe054bdd9527636886c20e42cd100c1 Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.015074 4775 scope.go:117] "RemoveContainer" containerID="564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.023213 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64dc0036-294e-4f2b-9b9d-cd540c42e558-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.035773 4775 scope.go:117] "RemoveContainer" containerID="58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa" Nov 26 06:29:08 crc kubenswrapper[4775]: E1126 06:29:08.036177 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa\": container with ID starting with 58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa not found: ID does not exist" containerID="58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.036218 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa"} err="failed to get container status \"58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa\": rpc error: code = NotFound desc = could not find container \"58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa\": container with ID starting with 58fd6ce4b7cf13db602566d33c31a056f835c51629a74389855ed738029e05fa not found: ID does not exist" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.036245 4775 scope.go:117] "RemoveContainer" containerID="718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9" Nov 26 06:29:08 crc kubenswrapper[4775]: E1126 06:29:08.036523 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9\": container with ID starting with 718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9 not found: ID does not exist" containerID="718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.036564 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9"} err="failed to get container status \"718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9\": rpc error: code = NotFound desc = could not find container \"718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9\": container with ID starting with 718ea51c873ba5a84e2f792de264671b54dbc8c58b196cc49d886f8e5ddc76a9 not found: ID does not exist" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.036591 4775 scope.go:117] "RemoveContainer" containerID="564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80" Nov 26 06:29:08 crc kubenswrapper[4775]: E1126 06:29:08.036850 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80\": container with ID starting with 564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80 not found: ID does not exist" containerID="564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.036879 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80"} err="failed to get container status \"564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80\": rpc error: code = NotFound desc = could not find container \"564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80\": container with ID starting with 564b3091b90832650e96c68bb9dde00e041a3fb01341906c4445bc7a6dd56a80 not found: ID does not exist" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.091450 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sjc6t"] Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.097248 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sjc6t"] Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.102115 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hpsd2"] Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.105343 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hpsd2"] Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.761532 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" event={"ID":"a4c5e343-455f-43ad-9936-8b64224702bf","Type":"ContainerStarted","Data":"e2a12f9e8d1b31857646f5304a687fb17056432cc786c828ff5a492cd0f1a2be"} Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.761692 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" event={"ID":"a4c5e343-455f-43ad-9936-8b64224702bf","Type":"ContainerStarted","Data":"655247bd6140e4a06335ad8537904d593fe054bdd9527636886c20e42cd100c1"} Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.761870 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.765831 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" Nov 26 06:29:08 crc kubenswrapper[4775]: I1126 06:29:08.792616 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sh8wn" podStartSLOduration=1.792600873 podStartE2EDuration="1.792600873s" podCreationTimestamp="2025-11-26 06:29:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:29:08.777471035 +0000 UTC m=+232.138774987" watchObservedRunningTime="2025-11-26 06:29:08.792600873 +0000 UTC m=+232.153904825" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291246 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-td9vf"] Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291470 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291481 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291489 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291496 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291519 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291525 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291533 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291538 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291546 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291551 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291559 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291564 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291573 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291579 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="extract-utilities" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291584 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291590 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291599 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291604 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291610 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291616 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291628 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291633 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291641 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291658 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="extract-content" Nov 26 06:29:09 crc kubenswrapper[4775]: E1126 06:29:09.291666 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerName="marketplace-operator" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291672 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerName="marketplace-operator" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291775 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291782 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291793 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291802 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" containerName="marketplace-operator" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.291811 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e21090b1-0905-474e-9148-fd349044b039" containerName="registry-server" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.292435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.295242 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.321556 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-td9vf"] Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.335847 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bdf9960-b1db-4086-9492-1b656cbe6e1b" path="/var/lib/kubelet/pods/4bdf9960-b1db-4086-9492-1b656cbe6e1b/volumes" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.337265 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64dc0036-294e-4f2b-9b9d-cd540c42e558" path="/var/lib/kubelet/pods/64dc0036-294e-4f2b-9b9d-cd540c42e558/volumes" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.338133 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd691a9-d0e3-4241-892d-cf2ff35959c1-catalog-content\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.338201 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhk8t\" (UniqueName: \"kubernetes.io/projected/7dd691a9-d0e3-4241-892d-cf2ff35959c1-kube-api-access-xhk8t\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.338224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd691a9-d0e3-4241-892d-cf2ff35959c1-utilities\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.338585 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ea68cee-a3d6-4ff7-b668-d87599fa4efd" path="/var/lib/kubelet/pods/7ea68cee-a3d6-4ff7-b668-d87599fa4efd/volumes" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.341959 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e21090b1-0905-474e-9148-fd349044b039" path="/var/lib/kubelet/pods/e21090b1-0905-474e-9148-fd349044b039/volumes" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.342950 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1365993-a452-4e00-92e0-64e1c08ce1c2" path="/var/lib/kubelet/pods/f1365993-a452-4e00-92e0-64e1c08ce1c2/volumes" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.441018 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhk8t\" (UniqueName: \"kubernetes.io/projected/7dd691a9-d0e3-4241-892d-cf2ff35959c1-kube-api-access-xhk8t\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.441128 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd691a9-d0e3-4241-892d-cf2ff35959c1-utilities\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.441299 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd691a9-d0e3-4241-892d-cf2ff35959c1-catalog-content\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.441603 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7dd691a9-d0e3-4241-892d-cf2ff35959c1-utilities\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.441948 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7dd691a9-d0e3-4241-892d-cf2ff35959c1-catalog-content\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.465462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhk8t\" (UniqueName: \"kubernetes.io/projected/7dd691a9-d0e3-4241-892d-cf2ff35959c1-kube-api-access-xhk8t\") pod \"certified-operators-td9vf\" (UID: \"7dd691a9-d0e3-4241-892d-cf2ff35959c1\") " pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.498011 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2ntql"] Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.499923 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.509165 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.512257 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ntql"] Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.542136 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8e888d-5042-4841-8b5e-18efc0d09506-catalog-content\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.542179 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4lxk\" (UniqueName: \"kubernetes.io/projected/2c8e888d-5042-4841-8b5e-18efc0d09506-kube-api-access-w4lxk\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.542256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8e888d-5042-4841-8b5e-18efc0d09506-utilities\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.612851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.643232 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8e888d-5042-4841-8b5e-18efc0d09506-utilities\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.643313 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8e888d-5042-4841-8b5e-18efc0d09506-catalog-content\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.643345 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4lxk\" (UniqueName: \"kubernetes.io/projected/2c8e888d-5042-4841-8b5e-18efc0d09506-kube-api-access-w4lxk\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.643960 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8e888d-5042-4841-8b5e-18efc0d09506-utilities\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.644018 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8e888d-5042-4841-8b5e-18efc0d09506-catalog-content\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.664467 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4lxk\" (UniqueName: \"kubernetes.io/projected/2c8e888d-5042-4841-8b5e-18efc0d09506-kube-api-access-w4lxk\") pod \"redhat-marketplace-2ntql\" (UID: \"2c8e888d-5042-4841-8b5e-18efc0d09506\") " pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.821375 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:09 crc kubenswrapper[4775]: I1126 06:29:09.997777 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-td9vf"] Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.024081 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2ntql"] Nov 26 06:29:10 crc kubenswrapper[4775]: W1126 06:29:10.028798 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c8e888d_5042_4841_8b5e_18efc0d09506.slice/crio-f1c70c2755db7269c28ddde072956f680f1bf56178d66ef23061254069c23c37 WatchSource:0}: Error finding container f1c70c2755db7269c28ddde072956f680f1bf56178d66ef23061254069c23c37: Status 404 returned error can't find the container with id f1c70c2755db7269c28ddde072956f680f1bf56178d66ef23061254069c23c37 Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.773373 4775 generic.go:334] "Generic (PLEG): container finished" podID="2c8e888d-5042-4841-8b5e-18efc0d09506" containerID="8021bd290d52322d37d79038c3f51df6e73905b6a816b21e64c180ca90d90963" exitCode=0 Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.773433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ntql" event={"ID":"2c8e888d-5042-4841-8b5e-18efc0d09506","Type":"ContainerDied","Data":"8021bd290d52322d37d79038c3f51df6e73905b6a816b21e64c180ca90d90963"} Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.773456 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ntql" event={"ID":"2c8e888d-5042-4841-8b5e-18efc0d09506","Type":"ContainerStarted","Data":"f1c70c2755db7269c28ddde072956f680f1bf56178d66ef23061254069c23c37"} Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.775814 4775 generic.go:334] "Generic (PLEG): container finished" podID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" containerID="fc345d8be427753a724f346d94ad1b0a10c0d6a0d9e25371219e241d74a2575e" exitCode=0 Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.776037 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9vf" event={"ID":"7dd691a9-d0e3-4241-892d-cf2ff35959c1","Type":"ContainerDied","Data":"fc345d8be427753a724f346d94ad1b0a10c0d6a0d9e25371219e241d74a2575e"} Nov 26 06:29:10 crc kubenswrapper[4775]: I1126 06:29:10.776084 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9vf" event={"ID":"7dd691a9-d0e3-4241-892d-cf2ff35959c1","Type":"ContainerStarted","Data":"e823c97b9506dac59760f682f594617d8b7724716eff0023820e8db78979598d"} Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.694135 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mhsdl"] Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.695748 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.698330 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.701067 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhsdl"] Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.769310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267d77e2-35df-457c-a0f8-9e7120db903d-utilities\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.769353 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267d77e2-35df-457c-a0f8-9e7120db903d-catalog-content\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.769378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vgzd\" (UniqueName: \"kubernetes.io/projected/267d77e2-35df-457c-a0f8-9e7120db903d-kube-api-access-8vgzd\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.782041 4775 generic.go:334] "Generic (PLEG): container finished" podID="2c8e888d-5042-4841-8b5e-18efc0d09506" containerID="ee2545ead6332e9446c6d3e741ae59e45800c0f5e762d05e98adf7bbb17bfd16" exitCode=0 Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.782101 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ntql" event={"ID":"2c8e888d-5042-4841-8b5e-18efc0d09506","Type":"ContainerDied","Data":"ee2545ead6332e9446c6d3e741ae59e45800c0f5e762d05e98adf7bbb17bfd16"} Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.783827 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9vf" event={"ID":"7dd691a9-d0e3-4241-892d-cf2ff35959c1","Type":"ContainerStarted","Data":"4cf96e30c86677e09dd47f0f444deb400474d7bcae06db0e538eb637e3d783ce"} Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.870589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vgzd\" (UniqueName: \"kubernetes.io/projected/267d77e2-35df-457c-a0f8-9e7120db903d-kube-api-access-8vgzd\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.870999 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267d77e2-35df-457c-a0f8-9e7120db903d-utilities\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.871021 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267d77e2-35df-457c-a0f8-9e7120db903d-catalog-content\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.871421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/267d77e2-35df-457c-a0f8-9e7120db903d-catalog-content\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.871923 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/267d77e2-35df-457c-a0f8-9e7120db903d-utilities\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.899664 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n8bhs"] Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.901061 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.903275 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vgzd\" (UniqueName: \"kubernetes.io/projected/267d77e2-35df-457c-a0f8-9e7120db903d-kube-api-access-8vgzd\") pod \"community-operators-mhsdl\" (UID: \"267d77e2-35df-457c-a0f8-9e7120db903d\") " pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.903578 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.905569 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8bhs"] Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.972308 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1786ea5-82b6-4354-98f5-ae908718294a-catalog-content\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.972505 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhsw8\" (UniqueName: \"kubernetes.io/projected/f1786ea5-82b6-4354-98f5-ae908718294a-kube-api-access-mhsw8\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:11 crc kubenswrapper[4775]: I1126 06:29:11.972620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1786ea5-82b6-4354-98f5-ae908718294a-utilities\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.034403 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.073298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1786ea5-82b6-4354-98f5-ae908718294a-catalog-content\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.073365 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhsw8\" (UniqueName: \"kubernetes.io/projected/f1786ea5-82b6-4354-98f5-ae908718294a-kube-api-access-mhsw8\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.073407 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1786ea5-82b6-4354-98f5-ae908718294a-utilities\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.074278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1786ea5-82b6-4354-98f5-ae908718294a-catalog-content\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.074353 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1786ea5-82b6-4354-98f5-ae908718294a-utilities\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.096155 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhsw8\" (UniqueName: \"kubernetes.io/projected/f1786ea5-82b6-4354-98f5-ae908718294a-kube-api-access-mhsw8\") pod \"redhat-operators-n8bhs\" (UID: \"f1786ea5-82b6-4354-98f5-ae908718294a\") " pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.246121 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.266097 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhsdl"] Nov 26 06:29:12 crc kubenswrapper[4775]: W1126 06:29:12.271424 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod267d77e2_35df_457c_a0f8_9e7120db903d.slice/crio-4486ded46e7a4ae7f3558759a75014c48a3f777a6ccc5a7d6bd2bf1df9d7fb17 WatchSource:0}: Error finding container 4486ded46e7a4ae7f3558759a75014c48a3f777a6ccc5a7d6bd2bf1df9d7fb17: Status 404 returned error can't find the container with id 4486ded46e7a4ae7f3558759a75014c48a3f777a6ccc5a7d6bd2bf1df9d7fb17 Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.405849 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8bhs"] Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.790846 4775 generic.go:334] "Generic (PLEG): container finished" podID="267d77e2-35df-457c-a0f8-9e7120db903d" containerID="233bc73b6ffe8eda63735b5dab9d8ef4e6d991c6d768e07267f440c904ed2818" exitCode=0 Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.790894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhsdl" event={"ID":"267d77e2-35df-457c-a0f8-9e7120db903d","Type":"ContainerDied","Data":"233bc73b6ffe8eda63735b5dab9d8ef4e6d991c6d768e07267f440c904ed2818"} Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.791152 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhsdl" event={"ID":"267d77e2-35df-457c-a0f8-9e7120db903d","Type":"ContainerStarted","Data":"4486ded46e7a4ae7f3558759a75014c48a3f777a6ccc5a7d6bd2bf1df9d7fb17"} Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.794518 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2ntql" event={"ID":"2c8e888d-5042-4841-8b5e-18efc0d09506","Type":"ContainerStarted","Data":"f13419253080ad3ab8ed51433ae673f350fb18786297bcbfedf6ee0f5b152000"} Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.797054 4775 generic.go:334] "Generic (PLEG): container finished" podID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" containerID="4cf96e30c86677e09dd47f0f444deb400474d7bcae06db0e538eb637e3d783ce" exitCode=0 Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.797283 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9vf" event={"ID":"7dd691a9-d0e3-4241-892d-cf2ff35959c1","Type":"ContainerDied","Data":"4cf96e30c86677e09dd47f0f444deb400474d7bcae06db0e538eb637e3d783ce"} Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.799703 4775 generic.go:334] "Generic (PLEG): container finished" podID="f1786ea5-82b6-4354-98f5-ae908718294a" containerID="4888144875de3262f274d00fb377fd529e1f2ffc3efdf2611220f0719e37578d" exitCode=0 Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.799755 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8bhs" event={"ID":"f1786ea5-82b6-4354-98f5-ae908718294a","Type":"ContainerDied","Data":"4888144875de3262f274d00fb377fd529e1f2ffc3efdf2611220f0719e37578d"} Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.799781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8bhs" event={"ID":"f1786ea5-82b6-4354-98f5-ae908718294a","Type":"ContainerStarted","Data":"d0e45f0e921bfe1c71ee4713ec65aa88b499ecd5b9783f5ecd991f948659e201"} Nov 26 06:29:12 crc kubenswrapper[4775]: I1126 06:29:12.837797 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2ntql" podStartSLOduration=2.215619448 podStartE2EDuration="3.83776232s" podCreationTimestamp="2025-11-26 06:29:09 +0000 UTC" firstStartedPulling="2025-11-26 06:29:10.775067339 +0000 UTC m=+234.136371331" lastFinishedPulling="2025-11-26 06:29:12.397210251 +0000 UTC m=+235.758514203" observedRunningTime="2025-11-26 06:29:12.834020957 +0000 UTC m=+236.195324909" watchObservedRunningTime="2025-11-26 06:29:12.83776232 +0000 UTC m=+236.199066312" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.692327 4775 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.694041 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.694746 4775 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.694947 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.695117 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66" gracePeriod=15 Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.695161 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6" gracePeriod=15 Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.695232 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3" gracePeriod=15 Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.695313 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a" gracePeriod=15 Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.695551 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c" gracePeriod=15 Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.695780 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.696131 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.696213 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.696287 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.696358 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.696465 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.696812 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.696890 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697003 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.697077 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697144 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.697217 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697283 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697452 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697532 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697600 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697687 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.697784 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.698018 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.698106 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.698300 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.703914 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.703982 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.704010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.704048 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.704094 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805239 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805301 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805341 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805370 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805368 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805393 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805450 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805458 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805704 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805844 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.805996 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.811667 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-td9vf" event={"ID":"7dd691a9-d0e3-4241-892d-cf2ff35959c1","Type":"ContainerStarted","Data":"b99ba3434da6c7b911e5b1dc1b0794d920b5e70b7debbe0aed87458be5da8373"} Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.812392 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.812692 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.907990 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.908090 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.908130 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.908179 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.908215 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: I1126 06:29:13.908254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:13 crc kubenswrapper[4775]: E1126 06:29:13.952986 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-mhsdl.187b7aab02e80777 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-mhsdl,UID:267d77e2-35df-457c-a0f8-9e7120db903d,APIVersion:v1,ResourceVersion:29511,FieldPath:spec.initContainers{extract-content},},Reason:Created,Message:Created container extract-content,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 06:29:13.952438135 +0000 UTC m=+237.313742087,LastTimestamp:2025-11-26 06:29:13.952438135 +0000 UTC m=+237.313742087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.817478 4775 generic.go:334] "Generic (PLEG): container finished" podID="267d77e2-35df-457c-a0f8-9e7120db903d" containerID="d6e4c98e37be6821b477aaed254dbfb5795d6f065285526ee32aa7eb90948a95" exitCode=0 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.817570 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhsdl" event={"ID":"267d77e2-35df-457c-a0f8-9e7120db903d","Type":"ContainerDied","Data":"d6e4c98e37be6821b477aaed254dbfb5795d6f065285526ee32aa7eb90948a95"} Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.818299 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.818611 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.818986 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.821110 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.823366 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.824190 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6" exitCode=0 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.824234 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a" exitCode=0 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.824251 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c" exitCode=0 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.824272 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3" exitCode=2 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.824275 4775 scope.go:117] "RemoveContainer" containerID="81bf02d75cce4471869779742accab7d2957529fd119af7a9495736048a07aae" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.826569 4775 generic.go:334] "Generic (PLEG): container finished" podID="228b00f8-984b-43be-a5ec-99e8414ee44c" containerID="a1562b1d5d21fe36ca9d15f331fc392945ca57e6727138a578799eb092cfc8a8" exitCode=0 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.826685 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"228b00f8-984b-43be-a5ec-99e8414ee44c","Type":"ContainerDied","Data":"a1562b1d5d21fe36ca9d15f331fc392945ca57e6727138a578799eb092cfc8a8"} Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.827437 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.827671 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.827953 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.828296 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.830169 4775 generic.go:334] "Generic (PLEG): container finished" podID="f1786ea5-82b6-4354-98f5-ae908718294a" containerID="6e9b003544a97eb68eb3a4e1066b869985d59d78bd2419cf63f4a9cacde5b1d5" exitCode=0 Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.830200 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8bhs" event={"ID":"f1786ea5-82b6-4354-98f5-ae908718294a","Type":"ContainerDied","Data":"6e9b003544a97eb68eb3a4e1066b869985d59d78bd2419cf63f4a9cacde5b1d5"} Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.830953 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.831428 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.834209 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.834605 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:14 crc kubenswrapper[4775]: I1126 06:29:14.835380 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:15 crc kubenswrapper[4775]: I1126 06:29:15.843191 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.532008 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.533016 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.533305 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.533542 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.533804 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.552349 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-var-lock\") pod \"228b00f8-984b-43be-a5ec-99e8414ee44c\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.552426 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-kubelet-dir\") pod \"228b00f8-984b-43be-a5ec-99e8414ee44c\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.552461 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228b00f8-984b-43be-a5ec-99e8414ee44c-kube-api-access\") pod \"228b00f8-984b-43be-a5ec-99e8414ee44c\" (UID: \"228b00f8-984b-43be-a5ec-99e8414ee44c\") " Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.555070 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "228b00f8-984b-43be-a5ec-99e8414ee44c" (UID: "228b00f8-984b-43be-a5ec-99e8414ee44c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.555098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-var-lock" (OuterVolumeSpecName: "var-lock") pod "228b00f8-984b-43be-a5ec-99e8414ee44c" (UID: "228b00f8-984b-43be-a5ec-99e8414ee44c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.558396 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/228b00f8-984b-43be-a5ec-99e8414ee44c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "228b00f8-984b-43be-a5ec-99e8414ee44c" (UID: "228b00f8-984b-43be-a5ec-99e8414ee44c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.614798 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.616464 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.616978 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.617348 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.617734 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.618047 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.618384 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654191 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654240 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654307 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654311 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654463 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654792 4775 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654832 4775 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-var-lock\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654860 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/228b00f8-984b-43be-a5ec-99e8414ee44c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654873 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228b00f8-984b-43be-a5ec-99e8414ee44c-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654890 4775 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.654950 4775 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.854185 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"228b00f8-984b-43be-a5ec-99e8414ee44c","Type":"ContainerDied","Data":"a4b17b6d6e119f4339b5d5c821cf11ea11bc25a107526fae1997e53d952c5643"} Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.854249 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b17b6d6e119f4339b5d5c821cf11ea11bc25a107526fae1997e53d952c5643" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.854210 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.856343 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.861384 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66" exitCode=0 Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.861438 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.861480 4775 scope.go:117] "RemoveContainer" containerID="d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.867192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8bhs" event={"ID":"f1786ea5-82b6-4354-98f5-ae908718294a","Type":"ContainerStarted","Data":"57b1b210c09b301d49d2129be27bfb078e9fade2ceefb0faf7bc08ac7c38a05d"} Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.869776 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.870784 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.871132 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.871419 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.871695 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.872244 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.872706 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.872830 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhsdl" event={"ID":"267d77e2-35df-457c-a0f8-9e7120db903d","Type":"ContainerStarted","Data":"2602192997bb839c0d68a41dbe3413e0aa3c3b9ca553a32ccdb101d3e0c6894d"} Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.873049 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.873948 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.874697 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.875106 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.875384 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.875577 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.875771 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.875911 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.876266 4775 scope.go:117] "RemoveContainer" containerID="105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.876676 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.878632 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.878910 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.879132 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.879355 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.890823 4775 scope.go:117] "RemoveContainer" containerID="df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.902326 4775 scope.go:117] "RemoveContainer" containerID="d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.916522 4775 scope.go:117] "RemoveContainer" containerID="377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.936963 4775 scope.go:117] "RemoveContainer" containerID="6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.964673 4775 scope.go:117] "RemoveContainer" containerID="d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6" Nov 26 06:29:16 crc kubenswrapper[4775]: E1126 06:29:16.965135 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\": container with ID starting with d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6 not found: ID does not exist" containerID="d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.965165 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6"} err="failed to get container status \"d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\": rpc error: code = NotFound desc = could not find container \"d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6\": container with ID starting with d51652491918f6b283f60ad725a090d70be699ea73acdcb9b1474ffbc2d143c6 not found: ID does not exist" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.965186 4775 scope.go:117] "RemoveContainer" containerID="105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a" Nov 26 06:29:16 crc kubenswrapper[4775]: E1126 06:29:16.965518 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\": container with ID starting with 105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a not found: ID does not exist" containerID="105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.965536 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a"} err="failed to get container status \"105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\": rpc error: code = NotFound desc = could not find container \"105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a\": container with ID starting with 105f4bbc9a5bf4c535413aa9a321a92c6e5ac42522305d36cb399b72b642127a not found: ID does not exist" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.965548 4775 scope.go:117] "RemoveContainer" containerID="df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c" Nov 26 06:29:16 crc kubenswrapper[4775]: E1126 06:29:16.965834 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\": container with ID starting with df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c not found: ID does not exist" containerID="df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.965857 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c"} err="failed to get container status \"df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\": rpc error: code = NotFound desc = could not find container \"df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c\": container with ID starting with df4b77d8738ade5de5e84973c52dc46ef6adacaf37328b368cc9277089c3730c not found: ID does not exist" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.965870 4775 scope.go:117] "RemoveContainer" containerID="d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3" Nov 26 06:29:16 crc kubenswrapper[4775]: E1126 06:29:16.966133 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\": container with ID starting with d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3 not found: ID does not exist" containerID="d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.966153 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3"} err="failed to get container status \"d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\": rpc error: code = NotFound desc = could not find container \"d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3\": container with ID starting with d42c46d55d71468a26642bdb225d64a5307f178f2361659c88a1e6385172c5d3 not found: ID does not exist" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.966164 4775 scope.go:117] "RemoveContainer" containerID="377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66" Nov 26 06:29:16 crc kubenswrapper[4775]: E1126 06:29:16.966453 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\": container with ID starting with 377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66 not found: ID does not exist" containerID="377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.966473 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66"} err="failed to get container status \"377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\": rpc error: code = NotFound desc = could not find container \"377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66\": container with ID starting with 377cc9e139d31cd709347649e0af7457e1040db061600ac31c7108fc4042be66 not found: ID does not exist" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.966485 4775 scope.go:117] "RemoveContainer" containerID="6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96" Nov 26 06:29:16 crc kubenswrapper[4775]: E1126 06:29:16.966747 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\": container with ID starting with 6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96 not found: ID does not exist" containerID="6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96" Nov 26 06:29:16 crc kubenswrapper[4775]: I1126 06:29:16.966771 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96"} err="failed to get container status \"6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\": rpc error: code = NotFound desc = could not find container \"6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96\": container with ID starting with 6f0c09172499b959048a18d04902883bd38a589a2bb6929d53b0a2ba013f8c96 not found: ID does not exist" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.329397 4775 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.330009 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.330200 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.330425 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.330820 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.338688 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.423807 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.424373 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.424737 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.424948 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.425915 4775 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:17 crc kubenswrapper[4775]: I1126 06:29:17.425974 4775 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.426497 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Nov 26 06:29:17 crc kubenswrapper[4775]: E1126 06:29:17.627329 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.028621 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.232419 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:29:18Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:29:18Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:29:18Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T06:29:18Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.232864 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.233287 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.233660 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.234040 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.234074 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.730815 4775 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:18 crc kubenswrapper[4775]: I1126 06:29:18.731487 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:18 crc kubenswrapper[4775]: W1126 06:29:18.758100 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b29e5da26676934ac9644a34dbc3bb310ef46f8b2089ff0f095d8ef92660ab67 WatchSource:0}: Error finding container b29e5da26676934ac9644a34dbc3bb310ef46f8b2089ff0f095d8ef92660ab67: Status 404 returned error can't find the container with id b29e5da26676934ac9644a34dbc3bb310ef46f8b2089ff0f095d8ef92660ab67 Nov 26 06:29:18 crc kubenswrapper[4775]: E1126 06:29:18.830086 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Nov 26 06:29:18 crc kubenswrapper[4775]: I1126 06:29:18.882792 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b29e5da26676934ac9644a34dbc3bb310ef46f8b2089ff0f095d8ef92660ab67"} Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.613062 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.613597 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.659966 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.660439 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.660625 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.660932 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.661259 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: E1126 06:29:19.732163 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{community-operators-mhsdl.187b7aab02e80777 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:community-operators-mhsdl,UID:267d77e2-35df-457c-a0f8-9e7120db903d,APIVersion:v1,ResourceVersion:29511,FieldPath:spec.initContainers{extract-content},},Reason:Created,Message:Created container extract-content,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 06:29:13.952438135 +0000 UTC m=+237.313742087,LastTimestamp:2025-11-26 06:29:13.952438135 +0000 UTC m=+237.313742087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.822480 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.822960 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.860606 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.861065 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.861435 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.861913 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.862178 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.862545 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.888629 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819"} Nov 26 06:29:19 crc kubenswrapper[4775]: E1126 06:29:19.889571 4775 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.889963 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.890555 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.891298 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.891780 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.892036 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.922513 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2ntql" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.923225 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.923514 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.923819 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.924244 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.924700 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.934702 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-td9vf" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.935243 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.935662 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.935999 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.936375 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:19 crc kubenswrapper[4775]: I1126 06:29:19.936658 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:20 crc kubenswrapper[4775]: E1126 06:29:20.431108 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Nov 26 06:29:20 crc kubenswrapper[4775]: E1126 06:29:20.898159 4775 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.035573 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.036011 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.092723 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.093264 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.093652 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.094097 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.094376 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.094646 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.245324 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.246902 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.292083 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.292672 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.293179 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.293534 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.293861 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.294213 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.978196 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mhsdl" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.979318 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.980142 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.980777 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.981378 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.982212 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.984020 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n8bhs" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.984850 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.985433 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.986118 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.986408 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:22 crc kubenswrapper[4775]: I1126 06:29:22.986813 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:23 crc kubenswrapper[4775]: E1126 06:29:23.330427 4775 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" volumeName="registry-storage" Nov 26 06:29:23 crc kubenswrapper[4775]: E1126 06:29:23.632441 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="6.4s" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.327320 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.329558 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.330178 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.330897 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.331581 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.332053 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.358140 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.358201 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:26 crc kubenswrapper[4775]: E1126 06:29:26.358887 4775 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.359625 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:26 crc kubenswrapper[4775]: W1126 06:29:26.392646 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-3125597dcd7fff4777249e6cbf68ef69684eb6256b96c88ebc733f0bb8ed52ab WatchSource:0}: Error finding container 3125597dcd7fff4777249e6cbf68ef69684eb6256b96c88ebc733f0bb8ed52ab: Status 404 returned error can't find the container with id 3125597dcd7fff4777249e6cbf68ef69684eb6256b96c88ebc733f0bb8ed52ab Nov 26 06:29:26 crc kubenswrapper[4775]: I1126 06:29:26.935093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3125597dcd7fff4777249e6cbf68ef69684eb6256b96c88ebc733f0bb8ed52ab"} Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.336128 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.336373 4775 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.336572 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.336773 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.336967 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.337125 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.940925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"eb0d20ffd535d2187ac63664d4dbfaef0f4aaaeb21fd6bde21b33d2ac4472d0a"} Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.944617 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.944651 4775 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db" exitCode=1 Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.944671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db"} Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.945178 4775 scope.go:117] "RemoveContainer" containerID="4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.945952 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.946116 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.946263 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.946535 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.946995 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.947225 4775 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:27 crc kubenswrapper[4775]: I1126 06:29:27.947411 4775 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.954167 4775 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="eb0d20ffd535d2187ac63664d4dbfaef0f4aaaeb21fd6bde21b33d2ac4472d0a" exitCode=0 Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.954305 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"eb0d20ffd535d2187ac63664d4dbfaef0f4aaaeb21fd6bde21b33d2ac4472d0a"} Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.954588 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.954622 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:28 crc kubenswrapper[4775]: E1126 06:29:28.955390 4775 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.955397 4775 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.956164 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.956762 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.957156 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.957547 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.957972 4775 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.958326 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.959308 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.959378 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d95e431edb8be74187af6f4406305da5af6e5e82ae07642865fe4547a47c499b"} Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.960216 4775 status_manager.go:851] "Failed to get status for pod" podUID="267d77e2-35df-457c-a0f8-9e7120db903d" pod="openshift-marketplace/community-operators-mhsdl" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-mhsdl\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.960445 4775 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.960756 4775 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.961113 4775 status_manager.go:851] "Failed to get status for pod" podUID="2c8e888d-5042-4841-8b5e-18efc0d09506" pod="openshift-marketplace/redhat-marketplace-2ntql" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-2ntql\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.961470 4775 status_manager.go:851] "Failed to get status for pod" podUID="f1786ea5-82b6-4354-98f5-ae908718294a" pod="openshift-marketplace/redhat-operators-n8bhs" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-n8bhs\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.961974 4775 status_manager.go:851] "Failed to get status for pod" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:28 crc kubenswrapper[4775]: I1126 06:29:28.962343 4775 status_manager.go:851] "Failed to get status for pod" podUID="7dd691a9-d0e3-4241-892d-cf2ff35959c1" pod="openshift-marketplace/certified-operators-td9vf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-td9vf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 06:29:29 crc kubenswrapper[4775]: I1126 06:29:29.970493 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"73f95b51f039fe2c87d659b9a034495027c785a4972b8569d9978b6d9da0f68a"} Nov 26 06:29:29 crc kubenswrapper[4775]: I1126 06:29:29.970753 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"aebb828b2e7d6088bb9b5e279228f272aa676db5beafc2c77b2ea8ad12a04d76"} Nov 26 06:29:29 crc kubenswrapper[4775]: I1126 06:29:29.970764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9fc50de24c1578b2911cb9aa40ad6aaf15f41efbea1e8f12a801deb63ad35586"} Nov 26 06:29:30 crc kubenswrapper[4775]: I1126 06:29:30.986363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"916dead5ce457745438661c0af0540a55e26bfa0460bd7c6d8378a24de0b6c27"} Nov 26 06:29:30 crc kubenswrapper[4775]: I1126 06:29:30.986639 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:30 crc kubenswrapper[4775]: I1126 06:29:30.986663 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:30 crc kubenswrapper[4775]: I1126 06:29:30.986761 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:30 crc kubenswrapper[4775]: I1126 06:29:30.986781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4ddfddf2a643fa896fd4a4e7a2f006cb499204553286a9ad11c8202a555f83fe"} Nov 26 06:29:31 crc kubenswrapper[4775]: I1126 06:29:31.360443 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:31 crc kubenswrapper[4775]: I1126 06:29:31.360521 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:31 crc kubenswrapper[4775]: I1126 06:29:31.368426 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]log ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]etcd ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/generic-apiserver-start-informers ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-filter ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-apiextensions-informers ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-apiextensions-controllers ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/crd-informer-synced ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-system-namespaces-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 26 06:29:31 crc kubenswrapper[4775]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 26 06:29:31 crc kubenswrapper[4775]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/bootstrap-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/start-kube-aggregator-informers ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-registration-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-discovery-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]autoregister-completion ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-openapi-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 26 06:29:31 crc kubenswrapper[4775]: livez check failed Nov 26 06:29:31 crc kubenswrapper[4775]: I1126 06:29:31.368483 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 06:29:31 crc kubenswrapper[4775]: I1126 06:29:31.888055 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:29:34 crc kubenswrapper[4775]: I1126 06:29:34.050767 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 26 06:29:34 crc kubenswrapper[4775]: I1126 06:29:34.051218 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 26 06:29:34 crc kubenswrapper[4775]: I1126 06:29:34.050164 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:29:36 crc kubenswrapper[4775]: I1126 06:29:36.001807 4775 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:36 crc kubenswrapper[4775]: I1126 06:29:36.023949 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:36 crc kubenswrapper[4775]: I1126 06:29:36.023981 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:36 crc kubenswrapper[4775]: I1126 06:29:36.074220 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="09495a12-703b-472b-974b-d644f8e2583f" Nov 26 06:29:44 crc kubenswrapper[4775]: I1126 06:29:44.050328 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 26 06:29:44 crc kubenswrapper[4775]: I1126 06:29:44.051103 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 26 06:29:45 crc kubenswrapper[4775]: I1126 06:29:45.321087 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 26 06:29:45 crc kubenswrapper[4775]: I1126 06:29:45.969504 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 26 06:29:46 crc kubenswrapper[4775]: I1126 06:29:46.282329 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 26 06:29:46 crc kubenswrapper[4775]: I1126 06:29:46.547476 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 26 06:29:46 crc kubenswrapper[4775]: I1126 06:29:46.713689 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 26 06:29:46 crc kubenswrapper[4775]: I1126 06:29:46.750218 4775 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 26 06:29:46 crc kubenswrapper[4775]: I1126 06:29:46.940028 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.133670 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.249099 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.249593 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.250284 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.258595 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.373451 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.426448 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.492370 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.516679 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.519209 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.524650 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.584363 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.661297 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.766438 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.767074 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.779294 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.781596 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.854058 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 26 06:29:47 crc kubenswrapper[4775]: I1126 06:29:47.996099 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.316991 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.459531 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.599549 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.647097 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.745017 4775 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.749466 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.751116 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.763871 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.786469 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.794928 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.848561 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 06:29:48 crc kubenswrapper[4775]: I1126 06:29:48.945410 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.013278 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.049042 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.085685 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.098054 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.155091 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.195050 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.239175 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.301467 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.369264 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.412240 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.423061 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.474978 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.546574 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.599003 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.629843 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.704562 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.724476 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.879075 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.891593 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.906107 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 26 06:29:49 crc kubenswrapper[4775]: I1126 06:29:49.950085 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.027365 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.077403 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.186380 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.250587 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.336337 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.373192 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.439445 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.525320 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.648878 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.686004 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.691064 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.700484 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.831536 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.892023 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.895888 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 06:29:50 crc kubenswrapper[4775]: I1126 06:29:50.929108 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.004951 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.011304 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.036887 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.057415 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.067309 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.126944 4775 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.168504 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.562431 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.611220 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.619471 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.664138 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.723681 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.734218 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.747134 4775 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.785637 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.810136 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.855026 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 26 06:29:51 crc kubenswrapper[4775]: I1126 06:29:51.917936 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.030314 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.068529 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.095603 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.121459 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.183507 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.227830 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.237540 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.254808 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.347139 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.376267 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.434390 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.443602 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.479360 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.523906 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.601052 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.609341 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.675595 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.825229 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.830698 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.842690 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.938985 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 06:29:52 crc kubenswrapper[4775]: I1126 06:29:52.967461 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.024190 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.027236 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.160085 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.290251 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.349309 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.366989 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.390355 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.403829 4775 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.407316 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n8bhs" podStartSLOduration=39.979574823 podStartE2EDuration="42.407296288s" podCreationTimestamp="2025-11-26 06:29:11 +0000 UTC" firstStartedPulling="2025-11-26 06:29:12.800821441 +0000 UTC m=+236.162125393" lastFinishedPulling="2025-11-26 06:29:15.228542906 +0000 UTC m=+238.589846858" observedRunningTime="2025-11-26 06:29:36.015778687 +0000 UTC m=+259.377082639" watchObservedRunningTime="2025-11-26 06:29:53.407296288 +0000 UTC m=+276.768600250" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.407890 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mhsdl" podStartSLOduration=39.966302846 podStartE2EDuration="42.407881154s" podCreationTimestamp="2025-11-26 06:29:11 +0000 UTC" firstStartedPulling="2025-11-26 06:29:12.792685236 +0000 UTC m=+236.153989188" lastFinishedPulling="2025-11-26 06:29:15.234263504 +0000 UTC m=+238.595567496" observedRunningTime="2025-11-26 06:29:36.071820964 +0000 UTC m=+259.433124936" watchObservedRunningTime="2025-11-26 06:29:53.407881154 +0000 UTC m=+276.769185116" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.409355 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-td9vf" podStartSLOduration=41.815234885 podStartE2EDuration="44.409346024s" podCreationTimestamp="2025-11-26 06:29:09 +0000 UTC" firstStartedPulling="2025-11-26 06:29:10.778308008 +0000 UTC m=+234.139611960" lastFinishedPulling="2025-11-26 06:29:13.372419147 +0000 UTC m=+236.733723099" observedRunningTime="2025-11-26 06:29:36.042884135 +0000 UTC m=+259.404188097" watchObservedRunningTime="2025-11-26 06:29:53.409346024 +0000 UTC m=+276.770649986" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.410494 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.410796 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.410933 4775 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.411368 4775 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="2940e7dc-6217-45e9-9aaf-b889adf5058a" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.415414 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.438577 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.438563318 podStartE2EDuration="17.438563318s" podCreationTimestamp="2025-11-26 06:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:29:53.437251972 +0000 UTC m=+276.798555934" watchObservedRunningTime="2025-11-26 06:29:53.438563318 +0000 UTC m=+276.799867270" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.475587 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.478341 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.493909 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.605878 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.636374 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.679959 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.689264 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.706073 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.726409 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.863150 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.870605 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.953525 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 26 06:29:53 crc kubenswrapper[4775]: I1126 06:29:53.957597 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.031625 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.051321 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.051412 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.051489 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.052592 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"d95e431edb8be74187af6f4406305da5af6e5e82ae07642865fe4547a47c499b"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.052851 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://d95e431edb8be74187af6f4406305da5af6e5e82ae07642865fe4547a47c499b" gracePeriod=30 Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.114031 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.204867 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.251900 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.329512 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.394963 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.583156 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.594620 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.626918 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.634524 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.645088 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 26 06:29:54 crc kubenswrapper[4775]: I1126 06:29:54.670494 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.012253 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.083862 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.092629 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.100949 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.107298 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.169954 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.210164 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.384530 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.462162 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.492607 4775 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.538815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.626264 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.738502 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.778410 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.788325 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.788994 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.796942 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.878883 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.911563 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 26 06:29:55 crc kubenswrapper[4775]: I1126 06:29:55.920180 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.029783 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.116156 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.129374 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.203531 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.209001 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.230204 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.334198 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.372227 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.373047 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.377645 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.380800 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.413134 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.426393 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.444463 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.448695 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.482788 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.560498 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.579505 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.618561 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.647473 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.649061 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.655888 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.702604 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.714414 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.719107 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.784160 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.822873 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.925604 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.941220 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.967146 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 26 06:29:56 crc kubenswrapper[4775]: I1126 06:29:56.979577 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.085321 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.131550 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.189354 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.190283 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.269103 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.291804 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.349125 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.351666 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.423339 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.476132 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.526539 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.540334 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.553561 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.624132 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.670807 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.690602 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.737710 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.789977 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.809243 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.882794 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.946132 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 26 06:29:57 crc kubenswrapper[4775]: I1126 06:29:57.972379 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.014200 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.017377 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.162564 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.223232 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.225058 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.234134 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.383289 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.420224 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.484414 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.596823 4775 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.597238 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819" gracePeriod=5 Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.771064 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 26 06:29:58 crc kubenswrapper[4775]: I1126 06:29:58.874825 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.006285 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.051887 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.335704 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.358641 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.388636 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.462126 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.644361 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.654383 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.912034 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 26 06:29:59 crc kubenswrapper[4775]: I1126 06:29:59.993344 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 26 06:30:00 crc kubenswrapper[4775]: I1126 06:30:00.381101 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 06:30:00 crc kubenswrapper[4775]: I1126 06:30:00.413901 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 26 06:30:01 crc kubenswrapper[4775]: I1126 06:30:01.243364 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 26 06:30:01 crc kubenswrapper[4775]: I1126 06:30:01.682787 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 26 06:30:01 crc kubenswrapper[4775]: I1126 06:30:01.833256 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 06:30:01 crc kubenswrapper[4775]: I1126 06:30:01.917244 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 06:30:02 crc kubenswrapper[4775]: I1126 06:30:02.068160 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 26 06:30:02 crc kubenswrapper[4775]: I1126 06:30:02.435883 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 26 06:30:02 crc kubenswrapper[4775]: I1126 06:30:02.464646 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 26 06:30:02 crc kubenswrapper[4775]: I1126 06:30:02.473347 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.180290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.180668 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.201334 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.201416 4775 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819" exitCode=137 Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.201488 4775 scope.go:117] "RemoveContainer" containerID="d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.201536 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.222672 4775 scope.go:117] "RemoveContainer" containerID="d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819" Nov 26 06:30:04 crc kubenswrapper[4775]: E1126 06:30:04.223308 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819\": container with ID starting with d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819 not found: ID does not exist" containerID="d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.223374 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819"} err="failed to get container status \"d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819\": rpc error: code = NotFound desc = could not find container \"d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819\": container with ID starting with d5362abee232360d17fad5a2b2c0bf85a2cfac8c856f580638e462c384d05819 not found: ID does not exist" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256230 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256322 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256372 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256449 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256594 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256594 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256650 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.256746 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.257070 4775 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.257124 4775 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.257148 4775 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.257173 4775 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.267352 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:30:04 crc kubenswrapper[4775]: I1126 06:30:04.358169 4775 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:05 crc kubenswrapper[4775]: I1126 06:30:05.338467 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 26 06:30:23 crc kubenswrapper[4775]: I1126 06:30:23.890663 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 06:30:24 crc kubenswrapper[4775]: I1126 06:30:24.352694 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 26 06:30:24 crc kubenswrapper[4775]: I1126 06:30:24.355196 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 06:30:24 crc kubenswrapper[4775]: I1126 06:30:24.355236 4775 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="d95e431edb8be74187af6f4406305da5af6e5e82ae07642865fe4547a47c499b" exitCode=137 Nov 26 06:30:24 crc kubenswrapper[4775]: I1126 06:30:24.355269 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"d95e431edb8be74187af6f4406305da5af6e5e82ae07642865fe4547a47c499b"} Nov 26 06:30:24 crc kubenswrapper[4775]: I1126 06:30:24.355304 4775 scope.go:117] "RemoveContainer" containerID="4b35f36fc6e69177b95db969cddf812501389143a116e71ee363138a192867db" Nov 26 06:30:25 crc kubenswrapper[4775]: I1126 06:30:25.368011 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Nov 26 06:30:25 crc kubenswrapper[4775]: I1126 06:30:25.369503 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"841d95030d7d66a9eb4be651930572ece25653f2bcca56b5b5b8e704ca2856e9"} Nov 26 06:30:29 crc kubenswrapper[4775]: I1126 06:30:29.163923 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 26 06:30:31 crc kubenswrapper[4775]: I1126 06:30:31.887368 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:30:34 crc kubenswrapper[4775]: I1126 06:30:34.050046 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:30:34 crc kubenswrapper[4775]: I1126 06:30:34.057091 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:30:34 crc kubenswrapper[4775]: I1126 06:30:34.440303 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.132765 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4"] Nov 26 06:30:46 crc kubenswrapper[4775]: E1126 06:30:46.133638 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.133657 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 06:30:46 crc kubenswrapper[4775]: E1126 06:30:46.133675 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" containerName="installer" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.133683 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" containerName="installer" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.133825 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="228b00f8-984b-43be-a5ec-99e8414ee44c" containerName="installer" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.133842 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.134284 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.135535 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.136565 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.186342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4"] Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.226281 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-dqwb7"] Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.226522 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" podUID="ab5f6914-f92e-47d6-9810-b61ce5151538" containerName="controller-manager" containerID="cri-o://5cf062586d6788f80095c119f8d8701c8e548e7944b9686b78677e97d455693c" gracePeriod=30 Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.229195 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs"] Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.229401 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" podUID="d05b88a6-499b-4c39-bb86-5629116aa323" containerName="route-controller-manager" containerID="cri-o://45f1d91905824ec0182e472333c57e9ea38390475ca09e3c4007a501545c4ef5" gracePeriod=30 Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.268419 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c54c2235-f554-4eda-bae7-b4fa7b68891e-secret-volume\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.268467 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c54c2235-f554-4eda-bae7-b4fa7b68891e-config-volume\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.268496 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpwxv\" (UniqueName: \"kubernetes.io/projected/c54c2235-f554-4eda-bae7-b4fa7b68891e-kube-api-access-tpwxv\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.369206 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c54c2235-f554-4eda-bae7-b4fa7b68891e-secret-volume\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.369246 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c54c2235-f554-4eda-bae7-b4fa7b68891e-config-volume\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.369273 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpwxv\" (UniqueName: \"kubernetes.io/projected/c54c2235-f554-4eda-bae7-b4fa7b68891e-kube-api-access-tpwxv\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.370513 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c54c2235-f554-4eda-bae7-b4fa7b68891e-config-volume\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.381473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c54c2235-f554-4eda-bae7-b4fa7b68891e-secret-volume\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.390140 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpwxv\" (UniqueName: \"kubernetes.io/projected/c54c2235-f554-4eda-bae7-b4fa7b68891e-kube-api-access-tpwxv\") pod \"collect-profiles-29402310-mhmw4\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.452600 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.575205 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab5f6914-f92e-47d6-9810-b61ce5151538" containerID="5cf062586d6788f80095c119f8d8701c8e548e7944b9686b78677e97d455693c" exitCode=0 Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.575294 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" event={"ID":"ab5f6914-f92e-47d6-9810-b61ce5151538","Type":"ContainerDied","Data":"5cf062586d6788f80095c119f8d8701c8e548e7944b9686b78677e97d455693c"} Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.578964 4775 generic.go:334] "Generic (PLEG): container finished" podID="d05b88a6-499b-4c39-bb86-5629116aa323" containerID="45f1d91905824ec0182e472333c57e9ea38390475ca09e3c4007a501545c4ef5" exitCode=0 Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.578994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" event={"ID":"d05b88a6-499b-4c39-bb86-5629116aa323","Type":"ContainerDied","Data":"45f1d91905824ec0182e472333c57e9ea38390475ca09e3c4007a501545c4ef5"} Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.579583 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.622673 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.772948 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-client-ca\") pod \"ab5f6914-f92e-47d6-9810-b61ce5151538\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.772997 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdrwx\" (UniqueName: \"kubernetes.io/projected/ab5f6914-f92e-47d6-9810-b61ce5151538-kube-api-access-sdrwx\") pod \"ab5f6914-f92e-47d6-9810-b61ce5151538\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d05b88a6-499b-4c39-bb86-5629116aa323-serving-cert\") pod \"d05b88a6-499b-4c39-bb86-5629116aa323\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773066 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-proxy-ca-bundles\") pod \"ab5f6914-f92e-47d6-9810-b61ce5151538\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773107 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5f6914-f92e-47d6-9810-b61ce5151538-serving-cert\") pod \"ab5f6914-f92e-47d6-9810-b61ce5151538\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773150 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82ht6\" (UniqueName: \"kubernetes.io/projected/d05b88a6-499b-4c39-bb86-5629116aa323-kube-api-access-82ht6\") pod \"d05b88a6-499b-4c39-bb86-5629116aa323\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773177 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-config\") pod \"d05b88a6-499b-4c39-bb86-5629116aa323\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773200 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-config\") pod \"ab5f6914-f92e-47d6-9810-b61ce5151538\" (UID: \"ab5f6914-f92e-47d6-9810-b61ce5151538\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.773215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-client-ca\") pod \"d05b88a6-499b-4c39-bb86-5629116aa323\" (UID: \"d05b88a6-499b-4c39-bb86-5629116aa323\") " Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.775222 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-config" (OuterVolumeSpecName: "config") pod "d05b88a6-499b-4c39-bb86-5629116aa323" (UID: "d05b88a6-499b-4c39-bb86-5629116aa323"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.775229 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-client-ca" (OuterVolumeSpecName: "client-ca") pod "ab5f6914-f92e-47d6-9810-b61ce5151538" (UID: "ab5f6914-f92e-47d6-9810-b61ce5151538"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.775261 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-client-ca" (OuterVolumeSpecName: "client-ca") pod "d05b88a6-499b-4c39-bb86-5629116aa323" (UID: "d05b88a6-499b-4c39-bb86-5629116aa323"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.775601 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-config" (OuterVolumeSpecName: "config") pod "ab5f6914-f92e-47d6-9810-b61ce5151538" (UID: "ab5f6914-f92e-47d6-9810-b61ce5151538"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.775860 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ab5f6914-f92e-47d6-9810-b61ce5151538" (UID: "ab5f6914-f92e-47d6-9810-b61ce5151538"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.779054 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05b88a6-499b-4c39-bb86-5629116aa323-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d05b88a6-499b-4c39-bb86-5629116aa323" (UID: "d05b88a6-499b-4c39-bb86-5629116aa323"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.779148 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5f6914-f92e-47d6-9810-b61ce5151538-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ab5f6914-f92e-47d6-9810-b61ce5151538" (UID: "ab5f6914-f92e-47d6-9810-b61ce5151538"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.779213 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab5f6914-f92e-47d6-9810-b61ce5151538-kube-api-access-sdrwx" (OuterVolumeSpecName: "kube-api-access-sdrwx") pod "ab5f6914-f92e-47d6-9810-b61ce5151538" (UID: "ab5f6914-f92e-47d6-9810-b61ce5151538"). InnerVolumeSpecName "kube-api-access-sdrwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.779263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05b88a6-499b-4c39-bb86-5629116aa323-kube-api-access-82ht6" (OuterVolumeSpecName: "kube-api-access-82ht6") pod "d05b88a6-499b-4c39-bb86-5629116aa323" (UID: "d05b88a6-499b-4c39-bb86-5629116aa323"). InnerVolumeSpecName "kube-api-access-82ht6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.874928 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab5f6914-f92e-47d6-9810-b61ce5151538-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875005 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82ht6\" (UniqueName: \"kubernetes.io/projected/d05b88a6-499b-4c39-bb86-5629116aa323-kube-api-access-82ht6\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875038 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875062 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875086 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d05b88a6-499b-4c39-bb86-5629116aa323-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875109 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875133 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdrwx\" (UniqueName: \"kubernetes.io/projected/ab5f6914-f92e-47d6-9810-b61ce5151538-kube-api-access-sdrwx\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875156 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d05b88a6-499b-4c39-bb86-5629116aa323-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.875180 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab5f6914-f92e-47d6-9810-b61ce5151538-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:46 crc kubenswrapper[4775]: I1126 06:30:46.902020 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.488403 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q"] Nov 26 06:30:47 crc kubenswrapper[4775]: E1126 06:30:47.488701 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5f6914-f92e-47d6-9810-b61ce5151538" containerName="controller-manager" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.488749 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5f6914-f92e-47d6-9810-b61ce5151538" containerName="controller-manager" Nov 26 06:30:47 crc kubenswrapper[4775]: E1126 06:30:47.488781 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05b88a6-499b-4c39-bb86-5629116aa323" containerName="route-controller-manager" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.488793 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05b88a6-499b-4c39-bb86-5629116aa323" containerName="route-controller-manager" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.488938 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5f6914-f92e-47d6-9810-b61ce5151538" containerName="controller-manager" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.488960 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05b88a6-499b-4c39-bb86-5629116aa323" containerName="route-controller-manager" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.489470 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.495024 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cc8d77586-4clpj"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.495890 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.500342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.505212 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cc8d77586-4clpj"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.585817 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.585813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs" event={"ID":"d05b88a6-499b-4c39-bb86-5629116aa323","Type":"ContainerDied","Data":"cd7bb8b7d387818d91504eac98d002cb4e889e53e07b00839efcd65c2b995a8d"} Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.586028 4775 scope.go:117] "RemoveContainer" containerID="45f1d91905824ec0182e472333c57e9ea38390475ca09e3c4007a501545c4ef5" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.587551 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.587485 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-dqwb7" event={"ID":"ab5f6914-f92e-47d6-9810-b61ce5151538","Type":"ContainerDied","Data":"a59a9a6766290842991eb3a0047b312e59a57531830cb960d435ba83d4285ccc"} Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.591679 4775 generic.go:334] "Generic (PLEG): container finished" podID="c54c2235-f554-4eda-bae7-b4fa7b68891e" containerID="a8945445912302fd8e3eba0af31209a72a6a00ed0850130cd144f9c71fc83233" exitCode=0 Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.591740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" event={"ID":"c54c2235-f554-4eda-bae7-b4fa7b68891e","Type":"ContainerDied","Data":"a8945445912302fd8e3eba0af31209a72a6a00ed0850130cd144f9c71fc83233"} Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.591767 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" event={"ID":"c54c2235-f554-4eda-bae7-b4fa7b68891e","Type":"ContainerStarted","Data":"95117e9288e6148b5e261bcf6787ae90208df922ee9526474d15f05bd2d8bc5b"} Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.623318 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.629511 4775 scope.go:117] "RemoveContainer" containerID="5cf062586d6788f80095c119f8d8701c8e548e7944b9686b78677e97d455693c" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.647222 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-84sgs"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.653456 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-dqwb7"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.665779 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-dqwb7"] Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685478 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-config\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685534 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6441e64-4f7a-4b68-afd7-143d98aa2715-serving-cert\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685560 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-proxy-ca-bundles\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685672 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-client-ca\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2qkw\" (UniqueName: \"kubernetes.io/projected/d6441e64-4f7a-4b68-afd7-143d98aa2715-kube-api-access-m2qkw\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685871 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12900edb-68a4-4c1b-8fa0-8a9132661325-serving-cert\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qbcw\" (UniqueName: \"kubernetes.io/projected/12900edb-68a4-4c1b-8fa0-8a9132661325-kube-api-access-8qbcw\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.685991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-client-ca\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.686024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-config\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qbcw\" (UniqueName: \"kubernetes.io/projected/12900edb-68a4-4c1b-8fa0-8a9132661325-kube-api-access-8qbcw\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-client-ca\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787200 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-config\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787228 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-config\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6441e64-4f7a-4b68-afd7-143d98aa2715-serving-cert\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787269 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-proxy-ca-bundles\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-client-ca\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787310 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12900edb-68a4-4c1b-8fa0-8a9132661325-serving-cert\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.787324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2qkw\" (UniqueName: \"kubernetes.io/projected/d6441e64-4f7a-4b68-afd7-143d98aa2715-kube-api-access-m2qkw\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.789436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-config\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.789475 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-client-ca\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.790984 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-client-ca\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.792311 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-proxy-ca-bundles\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.792987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-config\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.794119 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6441e64-4f7a-4b68-afd7-143d98aa2715-serving-cert\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.794441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12900edb-68a4-4c1b-8fa0-8a9132661325-serving-cert\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.809071 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2qkw\" (UniqueName: \"kubernetes.io/projected/d6441e64-4f7a-4b68-afd7-143d98aa2715-kube-api-access-m2qkw\") pod \"route-controller-manager-79cb97567b-nnw5q\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.818061 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.818530 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qbcw\" (UniqueName: \"kubernetes.io/projected/12900edb-68a4-4c1b-8fa0-8a9132661325-kube-api-access-8qbcw\") pod \"controller-manager-cc8d77586-4clpj\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:47 crc kubenswrapper[4775]: I1126 06:30:47.832446 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.051840 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q"] Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.099768 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cc8d77586-4clpj"] Nov 26 06:30:48 crc kubenswrapper[4775]: W1126 06:30:48.104822 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12900edb_68a4_4c1b_8fa0_8a9132661325.slice/crio-04efae52fa9f10933a668bbbb9e384ee74b2f9154d1142c3326567c2d6c4070b WatchSource:0}: Error finding container 04efae52fa9f10933a668bbbb9e384ee74b2f9154d1142c3326567c2d6c4070b: Status 404 returned error can't find the container with id 04efae52fa9f10933a668bbbb9e384ee74b2f9154d1142c3326567c2d6c4070b Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.601375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" event={"ID":"12900edb-68a4-4c1b-8fa0-8a9132661325","Type":"ContainerStarted","Data":"4051a1dda93f85ff24534d040d245e89877b6b43675359ff835667ed9945f446"} Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.601776 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.603078 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" event={"ID":"12900edb-68a4-4c1b-8fa0-8a9132661325","Type":"ContainerStarted","Data":"04efae52fa9f10933a668bbbb9e384ee74b2f9154d1142c3326567c2d6c4070b"} Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.620263 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.629819 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" event={"ID":"d6441e64-4f7a-4b68-afd7-143d98aa2715","Type":"ContainerStarted","Data":"441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3"} Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.629875 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.629891 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" event={"ID":"d6441e64-4f7a-4b68-afd7-143d98aa2715","Type":"ContainerStarted","Data":"0004c6955b2c77799bee00626f942ace8ccb841481e5ee3f54a21c7a9f591726"} Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.644138 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.662848 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" podStartSLOduration=2.662830887 podStartE2EDuration="2.662830887s" podCreationTimestamp="2025-11-26 06:30:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:30:48.661159541 +0000 UTC m=+332.022463493" watchObservedRunningTime="2025-11-26 06:30:48.662830887 +0000 UTC m=+332.024134869" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.663619 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" podStartSLOduration=2.6636116789999997 podStartE2EDuration="2.663611679s" podCreationTimestamp="2025-11-26 06:30:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:30:48.633564322 +0000 UTC m=+331.994868294" watchObservedRunningTime="2025-11-26 06:30:48.663611679 +0000 UTC m=+332.024915651" Nov 26 06:30:48 crc kubenswrapper[4775]: I1126 06:30:48.899102 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.006249 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c54c2235-f554-4eda-bae7-b4fa7b68891e-secret-volume\") pod \"c54c2235-f554-4eda-bae7-b4fa7b68891e\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.006332 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c54c2235-f554-4eda-bae7-b4fa7b68891e-config-volume\") pod \"c54c2235-f554-4eda-bae7-b4fa7b68891e\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.006368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpwxv\" (UniqueName: \"kubernetes.io/projected/c54c2235-f554-4eda-bae7-b4fa7b68891e-kube-api-access-tpwxv\") pod \"c54c2235-f554-4eda-bae7-b4fa7b68891e\" (UID: \"c54c2235-f554-4eda-bae7-b4fa7b68891e\") " Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.007324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54c2235-f554-4eda-bae7-b4fa7b68891e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c54c2235-f554-4eda-bae7-b4fa7b68891e" (UID: "c54c2235-f554-4eda-bae7-b4fa7b68891e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.013016 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54c2235-f554-4eda-bae7-b4fa7b68891e-kube-api-access-tpwxv" (OuterVolumeSpecName: "kube-api-access-tpwxv") pod "c54c2235-f554-4eda-bae7-b4fa7b68891e" (UID: "c54c2235-f554-4eda-bae7-b4fa7b68891e"). InnerVolumeSpecName "kube-api-access-tpwxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.013064 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c54c2235-f554-4eda-bae7-b4fa7b68891e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c54c2235-f554-4eda-bae7-b4fa7b68891e" (UID: "c54c2235-f554-4eda-bae7-b4fa7b68891e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.107735 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c54c2235-f554-4eda-bae7-b4fa7b68891e-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.107768 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpwxv\" (UniqueName: \"kubernetes.io/projected/c54c2235-f554-4eda-bae7-b4fa7b68891e-kube-api-access-tpwxv\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.107779 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c54c2235-f554-4eda-bae7-b4fa7b68891e-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.333774 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab5f6914-f92e-47d6-9810-b61ce5151538" path="/var/lib/kubelet/pods/ab5f6914-f92e-47d6-9810-b61ce5151538/volumes" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.334460 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05b88a6-499b-4c39-bb86-5629116aa323" path="/var/lib/kubelet/pods/d05b88a6-499b-4c39-bb86-5629116aa323/volumes" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.634497 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.634950 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4" event={"ID":"c54c2235-f554-4eda-bae7-b4fa7b68891e","Type":"ContainerDied","Data":"95117e9288e6148b5e261bcf6787ae90208df922ee9526474d15f05bd2d8bc5b"} Nov 26 06:30:49 crc kubenswrapper[4775]: I1126 06:30:49.634970 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95117e9288e6148b5e261bcf6787ae90208df922ee9526474d15f05bd2d8bc5b" Nov 26 06:30:51 crc kubenswrapper[4775]: I1126 06:30:51.225282 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q"] Nov 26 06:30:51 crc kubenswrapper[4775]: I1126 06:30:51.420257 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:30:51 crc kubenswrapper[4775]: I1126 06:30:51.420323 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:30:51 crc kubenswrapper[4775]: I1126 06:30:51.648003 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" podUID="d6441e64-4f7a-4b68-afd7-143d98aa2715" containerName="route-controller-manager" containerID="cri-o://441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3" gracePeriod=30 Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.149981 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.264270 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-client-ca\") pod \"d6441e64-4f7a-4b68-afd7-143d98aa2715\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.264332 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6441e64-4f7a-4b68-afd7-143d98aa2715-serving-cert\") pod \"d6441e64-4f7a-4b68-afd7-143d98aa2715\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.264403 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2qkw\" (UniqueName: \"kubernetes.io/projected/d6441e64-4f7a-4b68-afd7-143d98aa2715-kube-api-access-m2qkw\") pod \"d6441e64-4f7a-4b68-afd7-143d98aa2715\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.264431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-config\") pod \"d6441e64-4f7a-4b68-afd7-143d98aa2715\" (UID: \"d6441e64-4f7a-4b68-afd7-143d98aa2715\") " Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.265858 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-client-ca" (OuterVolumeSpecName: "client-ca") pod "d6441e64-4f7a-4b68-afd7-143d98aa2715" (UID: "d6441e64-4f7a-4b68-afd7-143d98aa2715"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.266178 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-config" (OuterVolumeSpecName: "config") pod "d6441e64-4f7a-4b68-afd7-143d98aa2715" (UID: "d6441e64-4f7a-4b68-afd7-143d98aa2715"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.271477 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6441e64-4f7a-4b68-afd7-143d98aa2715-kube-api-access-m2qkw" (OuterVolumeSpecName: "kube-api-access-m2qkw") pod "d6441e64-4f7a-4b68-afd7-143d98aa2715" (UID: "d6441e64-4f7a-4b68-afd7-143d98aa2715"). InnerVolumeSpecName "kube-api-access-m2qkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.273314 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6441e64-4f7a-4b68-afd7-143d98aa2715-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d6441e64-4f7a-4b68-afd7-143d98aa2715" (UID: "d6441e64-4f7a-4b68-afd7-143d98aa2715"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.365969 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.366390 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2qkw\" (UniqueName: \"kubernetes.io/projected/d6441e64-4f7a-4b68-afd7-143d98aa2715-kube-api-access-m2qkw\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.366402 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d6441e64-4f7a-4b68-afd7-143d98aa2715-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.366412 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6441e64-4f7a-4b68-afd7-143d98aa2715-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.495640 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788767b658-9st8r"] Nov 26 06:30:52 crc kubenswrapper[4775]: E1126 06:30:52.495912 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6441e64-4f7a-4b68-afd7-143d98aa2715" containerName="route-controller-manager" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.495932 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6441e64-4f7a-4b68-afd7-143d98aa2715" containerName="route-controller-manager" Nov 26 06:30:52 crc kubenswrapper[4775]: E1126 06:30:52.495948 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54c2235-f554-4eda-bae7-b4fa7b68891e" containerName="collect-profiles" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.495960 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54c2235-f554-4eda-bae7-b4fa7b68891e" containerName="collect-profiles" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.496105 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54c2235-f554-4eda-bae7-b4fa7b68891e" containerName="collect-profiles" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.496127 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6441e64-4f7a-4b68-afd7-143d98aa2715" containerName="route-controller-manager" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.496825 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.512389 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788767b658-9st8r"] Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.658468 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6441e64-4f7a-4b68-afd7-143d98aa2715" containerID="441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3" exitCode=0 Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.658511 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" event={"ID":"d6441e64-4f7a-4b68-afd7-143d98aa2715","Type":"ContainerDied","Data":"441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3"} Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.658538 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" event={"ID":"d6441e64-4f7a-4b68-afd7-143d98aa2715","Type":"ContainerDied","Data":"0004c6955b2c77799bee00626f942ace8ccb841481e5ee3f54a21c7a9f591726"} Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.658553 4775 scope.go:117] "RemoveContainer" containerID="441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.659113 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.669950 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b27849-36cb-4b26-9df0-721571d2ac9f-config\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.670021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23b27849-36cb-4b26-9df0-721571d2ac9f-client-ca\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.670441 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlztc\" (UniqueName: \"kubernetes.io/projected/23b27849-36cb-4b26-9df0-721571d2ac9f-kube-api-access-wlztc\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.670635 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23b27849-36cb-4b26-9df0-721571d2ac9f-serving-cert\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.684127 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q"] Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.689497 4775 scope.go:117] "RemoveContainer" containerID="441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3" Nov 26 06:30:52 crc kubenswrapper[4775]: E1126 06:30:52.690268 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3\": container with ID starting with 441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3 not found: ID does not exist" containerID="441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.690315 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3"} err="failed to get container status \"441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3\": rpc error: code = NotFound desc = could not find container \"441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3\": container with ID starting with 441747c4fef066e6b3e063d4726393fe0c524cd04890486151483e18b25f8de3 not found: ID does not exist" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.694635 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-79cb97567b-nnw5q"] Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.772208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlztc\" (UniqueName: \"kubernetes.io/projected/23b27849-36cb-4b26-9df0-721571d2ac9f-kube-api-access-wlztc\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.772283 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23b27849-36cb-4b26-9df0-721571d2ac9f-serving-cert\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.772306 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b27849-36cb-4b26-9df0-721571d2ac9f-config\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.772367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23b27849-36cb-4b26-9df0-721571d2ac9f-client-ca\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.773305 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23b27849-36cb-4b26-9df0-721571d2ac9f-client-ca\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.776434 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b27849-36cb-4b26-9df0-721571d2ac9f-config\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.778335 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23b27849-36cb-4b26-9df0-721571d2ac9f-serving-cert\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.793577 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlztc\" (UniqueName: \"kubernetes.io/projected/23b27849-36cb-4b26-9df0-721571d2ac9f-kube-api-access-wlztc\") pod \"route-controller-manager-788767b658-9st8r\" (UID: \"23b27849-36cb-4b26-9df0-721571d2ac9f\") " pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:52 crc kubenswrapper[4775]: I1126 06:30:52.818360 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.227125 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-788767b658-9st8r"] Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.343611 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6441e64-4f7a-4b68-afd7-143d98aa2715" path="/var/lib/kubelet/pods/d6441e64-4f7a-4b68-afd7-143d98aa2715/volumes" Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.665977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" event={"ID":"23b27849-36cb-4b26-9df0-721571d2ac9f","Type":"ContainerStarted","Data":"92dfc0f795897149f575c94ce45f1996ff22911bdfefdfe82ee8ef47053c5eef"} Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.666444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.666466 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" event={"ID":"23b27849-36cb-4b26-9df0-721571d2ac9f","Type":"ContainerStarted","Data":"30ef0ef54f6a7009838f9583d5150cdbf1541e60d0be6351c37fa624f7343fc8"} Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.885221 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" Nov 26 06:30:53 crc kubenswrapper[4775]: I1126 06:30:53.908491 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-788767b658-9st8r" podStartSLOduration=2.908462503 podStartE2EDuration="2.908462503s" podCreationTimestamp="2025-11-26 06:30:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:30:53.694022234 +0000 UTC m=+337.055326216" watchObservedRunningTime="2025-11-26 06:30:53.908462503 +0000 UTC m=+337.269766495" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.230891 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cc8d77586-4clpj"] Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.231515 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" podUID="12900edb-68a4-4c1b-8fa0-8a9132661325" containerName="controller-manager" containerID="cri-o://4051a1dda93f85ff24534d040d245e89877b6b43675359ff835667ed9945f446" gracePeriod=30 Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.774670 4775 generic.go:334] "Generic (PLEG): container finished" podID="12900edb-68a4-4c1b-8fa0-8a9132661325" containerID="4051a1dda93f85ff24534d040d245e89877b6b43675359ff835667ed9945f446" exitCode=0 Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.774940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" event={"ID":"12900edb-68a4-4c1b-8fa0-8a9132661325","Type":"ContainerDied","Data":"4051a1dda93f85ff24534d040d245e89877b6b43675359ff835667ed9945f446"} Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.774968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" event={"ID":"12900edb-68a4-4c1b-8fa0-8a9132661325","Type":"ContainerDied","Data":"04efae52fa9f10933a668bbbb9e384ee74b2f9154d1142c3326567c2d6c4070b"} Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.774977 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04efae52fa9f10933a668bbbb9e384ee74b2f9154d1142c3326567c2d6c4070b" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.795682 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.856860 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-proxy-ca-bundles\") pod \"12900edb-68a4-4c1b-8fa0-8a9132661325\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.856921 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-client-ca\") pod \"12900edb-68a4-4c1b-8fa0-8a9132661325\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.856949 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-config\") pod \"12900edb-68a4-4c1b-8fa0-8a9132661325\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.856967 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12900edb-68a4-4c1b-8fa0-8a9132661325-serving-cert\") pod \"12900edb-68a4-4c1b-8fa0-8a9132661325\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.856999 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qbcw\" (UniqueName: \"kubernetes.io/projected/12900edb-68a4-4c1b-8fa0-8a9132661325-kube-api-access-8qbcw\") pod \"12900edb-68a4-4c1b-8fa0-8a9132661325\" (UID: \"12900edb-68a4-4c1b-8fa0-8a9132661325\") " Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.857577 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "12900edb-68a4-4c1b-8fa0-8a9132661325" (UID: "12900edb-68a4-4c1b-8fa0-8a9132661325"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.857686 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-client-ca" (OuterVolumeSpecName: "client-ca") pod "12900edb-68a4-4c1b-8fa0-8a9132661325" (UID: "12900edb-68a4-4c1b-8fa0-8a9132661325"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.857703 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-config" (OuterVolumeSpecName: "config") pod "12900edb-68a4-4c1b-8fa0-8a9132661325" (UID: "12900edb-68a4-4c1b-8fa0-8a9132661325"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.862911 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12900edb-68a4-4c1b-8fa0-8a9132661325-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "12900edb-68a4-4c1b-8fa0-8a9132661325" (UID: "12900edb-68a4-4c1b-8fa0-8a9132661325"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.862940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12900edb-68a4-4c1b-8fa0-8a9132661325-kube-api-access-8qbcw" (OuterVolumeSpecName: "kube-api-access-8qbcw") pod "12900edb-68a4-4c1b-8fa0-8a9132661325" (UID: "12900edb-68a4-4c1b-8fa0-8a9132661325"). InnerVolumeSpecName "kube-api-access-8qbcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.958420 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12900edb-68a4-4c1b-8fa0-8a9132661325-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.958454 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qbcw\" (UniqueName: \"kubernetes.io/projected/12900edb-68a4-4c1b-8fa0-8a9132661325-kube-api-access-8qbcw\") on node \"crc\" DevicePath \"\"" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.958466 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.958475 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:31:11 crc kubenswrapper[4775]: I1126 06:31:11.958483 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12900edb-68a4-4c1b-8fa0-8a9132661325-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.501322 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9"] Nov 26 06:31:12 crc kubenswrapper[4775]: E1126 06:31:12.501510 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12900edb-68a4-4c1b-8fa0-8a9132661325" containerName="controller-manager" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.501522 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="12900edb-68a4-4c1b-8fa0-8a9132661325" containerName="controller-manager" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.501609 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="12900edb-68a4-4c1b-8fa0-8a9132661325" containerName="controller-manager" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.501917 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.515187 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9"] Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.566541 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67hsc\" (UniqueName: \"kubernetes.io/projected/360d714c-d379-4b21-903f-731141a36e83-kube-api-access-67hsc\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.566617 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-client-ca\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.566786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360d714c-d379-4b21-903f-731141a36e83-serving-cert\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.566850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-config\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.566962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-proxy-ca-bundles\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.667669 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67hsc\" (UniqueName: \"kubernetes.io/projected/360d714c-d379-4b21-903f-731141a36e83-kube-api-access-67hsc\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.667736 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-client-ca\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.667783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360d714c-d379-4b21-903f-731141a36e83-serving-cert\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.667803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-config\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.667824 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-proxy-ca-bundles\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.668623 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-client-ca\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.668950 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-proxy-ca-bundles\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.669113 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360d714c-d379-4b21-903f-731141a36e83-config\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.681580 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360d714c-d379-4b21-903f-731141a36e83-serving-cert\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.689124 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67hsc\" (UniqueName: \"kubernetes.io/projected/360d714c-d379-4b21-903f-731141a36e83-kube-api-access-67hsc\") pod \"controller-manager-66f5bd6f8f-kggp9\" (UID: \"360d714c-d379-4b21-903f-731141a36e83\") " pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.779434 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc8d77586-4clpj" Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.808415 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cc8d77586-4clpj"] Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.811912 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-cc8d77586-4clpj"] Nov 26 06:31:12 crc kubenswrapper[4775]: I1126 06:31:12.823836 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.113257 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9"] Nov 26 06:31:13 crc kubenswrapper[4775]: W1126 06:31:13.117243 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod360d714c_d379_4b21_903f_731141a36e83.slice/crio-0408d489658434c4b733da330f68d5a4b6c5501eaa53060fe7d310c89700e6b4 WatchSource:0}: Error finding container 0408d489658434c4b733da330f68d5a4b6c5501eaa53060fe7d310c89700e6b4: Status 404 returned error can't find the container with id 0408d489658434c4b733da330f68d5a4b6c5501eaa53060fe7d310c89700e6b4 Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.338901 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12900edb-68a4-4c1b-8fa0-8a9132661325" path="/var/lib/kubelet/pods/12900edb-68a4-4c1b-8fa0-8a9132661325/volumes" Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.791680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" event={"ID":"360d714c-d379-4b21-903f-731141a36e83","Type":"ContainerStarted","Data":"1a3818b9456721cf0bfcbbcabcc4e0db9275c282e054d15a1f7fde18c12996a5"} Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.791972 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" event={"ID":"360d714c-d379-4b21-903f-731141a36e83","Type":"ContainerStarted","Data":"0408d489658434c4b733da330f68d5a4b6c5501eaa53060fe7d310c89700e6b4"} Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.793212 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.798974 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" Nov 26 06:31:13 crc kubenswrapper[4775]: I1126 06:31:13.828996 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66f5bd6f8f-kggp9" podStartSLOduration=2.828980396 podStartE2EDuration="2.828980396s" podCreationTimestamp="2025-11-26 06:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:31:13.814503068 +0000 UTC m=+357.175807040" watchObservedRunningTime="2025-11-26 06:31:13.828980396 +0000 UTC m=+357.190284348" Nov 26 06:31:21 crc kubenswrapper[4775]: I1126 06:31:21.420224 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:31:21 crc kubenswrapper[4775]: I1126 06:31:21.421055 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.746031 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fmdbw"] Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.746863 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.764284 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fmdbw"] Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904543 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904662 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-trusted-ca\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904685 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-bound-sa-token\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904732 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904753 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-registry-tls\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904773 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-844cl\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-kube-api-access-844cl\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.904800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-registry-certificates\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:22 crc kubenswrapper[4775]: I1126 06:31:22.937869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-trusted-ca\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006103 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-bound-sa-token\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-registry-tls\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006197 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-844cl\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-kube-api-access-844cl\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-registry-certificates\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.006393 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.007147 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.008300 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-trusted-ca\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.008543 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-registry-certificates\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.014075 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-registry-tls\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.014083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.024630 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-844cl\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-kube-api-access-844cl\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.034030 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/86f06c4d-fb94-48ad-9b5d-0c3745b36c41-bound-sa-token\") pod \"image-registry-66df7c8f76-fmdbw\" (UID: \"86f06c4d-fb94-48ad-9b5d-0c3745b36c41\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.078790 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.563372 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fmdbw"] Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.855679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" event={"ID":"86f06c4d-fb94-48ad-9b5d-0c3745b36c41","Type":"ContainerStarted","Data":"727e0ee7c0055a3512d7c3c4aec2aa4cd217901ca90430cf38eda5e2acf86109"} Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.855780 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" event={"ID":"86f06c4d-fb94-48ad-9b5d-0c3745b36c41","Type":"ContainerStarted","Data":"1619104446b17563fea7a6e8068a414e958e7f1e548886d6bf0d15428d0e1adc"} Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.855917 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:23 crc kubenswrapper[4775]: I1126 06:31:23.889593 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" podStartSLOduration=1.889567972 podStartE2EDuration="1.889567972s" podCreationTimestamp="2025-11-26 06:31:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:31:23.883209787 +0000 UTC m=+367.244513769" watchObservedRunningTime="2025-11-26 06:31:23.889567972 +0000 UTC m=+367.250871954" Nov 26 06:31:43 crc kubenswrapper[4775]: I1126 06:31:43.085094 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fmdbw" Nov 26 06:31:43 crc kubenswrapper[4775]: I1126 06:31:43.160591 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4h8zh"] Nov 26 06:31:51 crc kubenswrapper[4775]: I1126 06:31:51.419916 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:31:51 crc kubenswrapper[4775]: I1126 06:31:51.421913 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:31:51 crc kubenswrapper[4775]: I1126 06:31:51.422006 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:31:51 crc kubenswrapper[4775]: I1126 06:31:51.422990 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a22b92d95ea00f90fd226634f524ff08e52cb46e50050a0f9a3125f6b85b48ba"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:31:51 crc kubenswrapper[4775]: I1126 06:31:51.423128 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://a22b92d95ea00f90fd226634f524ff08e52cb46e50050a0f9a3125f6b85b48ba" gracePeriod=600 Nov 26 06:31:52 crc kubenswrapper[4775]: I1126 06:31:52.046570 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="a22b92d95ea00f90fd226634f524ff08e52cb46e50050a0f9a3125f6b85b48ba" exitCode=0 Nov 26 06:31:52 crc kubenswrapper[4775]: I1126 06:31:52.046759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"a22b92d95ea00f90fd226634f524ff08e52cb46e50050a0f9a3125f6b85b48ba"} Nov 26 06:31:52 crc kubenswrapper[4775]: I1126 06:31:52.047039 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"988eb2b1e284ef66703f1f37b2da654fd669430dce2be08f52d009293283f65e"} Nov 26 06:31:52 crc kubenswrapper[4775]: I1126 06:31:52.047082 4775 scope.go:117] "RemoveContainer" containerID="f6638514765c36e22af89fee7dceb5c8230cf951685685a33ad0e558b6365d5f" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.224024 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" podUID="088e2d8e-dfb4-4a5d-b681-d7d46573b629" containerName="registry" containerID="cri-o://216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f" gracePeriod=30 Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.688596 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.832005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-trusted-ca\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.832367 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.832419 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-certificates\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.832442 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/088e2d8e-dfb4-4a5d-b681-d7d46573b629-installation-pull-secrets\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.833255 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.833300 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.833336 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/088e2d8e-dfb4-4a5d-b681-d7d46573b629-ca-trust-extracted\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.833519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82hfb\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-kube-api-access-82hfb\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.833557 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-bound-sa-token\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.833579 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-tls\") pod \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\" (UID: \"088e2d8e-dfb4-4a5d-b681-d7d46573b629\") " Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.834245 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.834282 4775 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.839762 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/088e2d8e-dfb4-4a5d-b681-d7d46573b629-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.840555 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.840940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.841459 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.843048 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-kube-api-access-82hfb" (OuterVolumeSpecName: "kube-api-access-82hfb") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "kube-api-access-82hfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.852729 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/088e2d8e-dfb4-4a5d-b681-d7d46573b629-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "088e2d8e-dfb4-4a5d-b681-d7d46573b629" (UID: "088e2d8e-dfb4-4a5d-b681-d7d46573b629"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.935289 4775 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/088e2d8e-dfb4-4a5d-b681-d7d46573b629-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.935357 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82hfb\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-kube-api-access-82hfb\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.935387 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.935413 4775 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/088e2d8e-dfb4-4a5d-b681-d7d46573b629-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:08 crc kubenswrapper[4775]: I1126 06:32:08.935437 4775 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/088e2d8e-dfb4-4a5d-b681-d7d46573b629-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.168309 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.168325 4775 generic.go:334] "Generic (PLEG): container finished" podID="088e2d8e-dfb4-4a5d-b681-d7d46573b629" containerID="216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f" exitCode=0 Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.168374 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" event={"ID":"088e2d8e-dfb4-4a5d-b681-d7d46573b629","Type":"ContainerDied","Data":"216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f"} Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.168505 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-4h8zh" event={"ID":"088e2d8e-dfb4-4a5d-b681-d7d46573b629","Type":"ContainerDied","Data":"33c7405d32376be90f51708bac779f6618da214bd757b0f905e3bf5ac9a9fead"} Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.168531 4775 scope.go:117] "RemoveContainer" containerID="216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f" Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.199924 4775 scope.go:117] "RemoveContainer" containerID="216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f" Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.201895 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4h8zh"] Nov 26 06:32:09 crc kubenswrapper[4775]: E1126 06:32:09.205033 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f\": container with ID starting with 216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f not found: ID does not exist" containerID="216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f" Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.205095 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f"} err="failed to get container status \"216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f\": rpc error: code = NotFound desc = could not find container \"216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f\": container with ID starting with 216f320681fd29b9dc281d08210bdc51e75ff80c17b184febb5197174685017f not found: ID does not exist" Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.208639 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-4h8zh"] Nov 26 06:32:09 crc kubenswrapper[4775]: I1126 06:32:09.345989 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088e2d8e-dfb4-4a5d-b681-d7d46573b629" path="/var/lib/kubelet/pods/088e2d8e-dfb4-4a5d-b681-d7d46573b629/volumes" Nov 26 06:33:51 crc kubenswrapper[4775]: I1126 06:33:51.420168 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:33:51 crc kubenswrapper[4775]: I1126 06:33:51.420847 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:34:17 crc kubenswrapper[4775]: I1126 06:34:17.564115 4775 scope.go:117] "RemoveContainer" containerID="edbb9e673d0832dd0cc4bbf4a0cf4d025acaf0fd9e247dfdec5fde2d46d042a1" Nov 26 06:34:17 crc kubenswrapper[4775]: I1126 06:34:17.600671 4775 scope.go:117] "RemoveContainer" containerID="a9523fcc45fcd882d1780521daf6d31b4eea97212430d7185c77443d1afb5db1" Nov 26 06:34:17 crc kubenswrapper[4775]: I1126 06:34:17.638459 4775 scope.go:117] "RemoveContainer" containerID="c73c84d3cff170e8666f0d22dcb395b56c19004f652c7d001ede6a8e488e12ee" Nov 26 06:34:21 crc kubenswrapper[4775]: I1126 06:34:21.419824 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:34:21 crc kubenswrapper[4775]: I1126 06:34:21.420205 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.909403 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-99ngt"] Nov 26 06:34:35 crc kubenswrapper[4775]: E1126 06:34:35.910647 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088e2d8e-dfb4-4a5d-b681-d7d46573b629" containerName="registry" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.910727 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="088e2d8e-dfb4-4a5d-b681-d7d46573b629" containerName="registry" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.910874 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="088e2d8e-dfb4-4a5d-b681-d7d46573b629" containerName="registry" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.911238 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.912901 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bh4vt" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.913276 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.913434 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.927824 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-99ngt"] Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.948261 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-f9chx"] Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.952575 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-f9chx" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.956331 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-pcnc6" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.956781 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-mn5gf"] Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.962772 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.968877 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vzkpf" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.978741 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-f9chx"] Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.981424 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-mn5gf"] Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.983218 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8wvf\" (UniqueName: \"kubernetes.io/projected/1d540303-134e-4ca8-9597-9eb8dda29ef4-kube-api-access-m8wvf\") pod \"cert-manager-cainjector-7f985d654d-99ngt\" (UID: \"1d540303-134e-4ca8-9597-9eb8dda29ef4\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.983273 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48rk7\" (UniqueName: \"kubernetes.io/projected/8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd-kube-api-access-48rk7\") pod \"cert-manager-webhook-5655c58dd6-mn5gf\" (UID: \"8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:35 crc kubenswrapper[4775]: I1126 06:34:35.983319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcds4\" (UniqueName: \"kubernetes.io/projected/4fe2d6c0-b093-452f-8dfa-0db642716f10-kube-api-access-kcds4\") pod \"cert-manager-5b446d88c5-f9chx\" (UID: \"4fe2d6c0-b093-452f-8dfa-0db642716f10\") " pod="cert-manager/cert-manager-5b446d88c5-f9chx" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.104916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8wvf\" (UniqueName: \"kubernetes.io/projected/1d540303-134e-4ca8-9597-9eb8dda29ef4-kube-api-access-m8wvf\") pod \"cert-manager-cainjector-7f985d654d-99ngt\" (UID: \"1d540303-134e-4ca8-9597-9eb8dda29ef4\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.104967 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48rk7\" (UniqueName: \"kubernetes.io/projected/8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd-kube-api-access-48rk7\") pod \"cert-manager-webhook-5655c58dd6-mn5gf\" (UID: \"8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.104993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcds4\" (UniqueName: \"kubernetes.io/projected/4fe2d6c0-b093-452f-8dfa-0db642716f10-kube-api-access-kcds4\") pod \"cert-manager-5b446d88c5-f9chx\" (UID: \"4fe2d6c0-b093-452f-8dfa-0db642716f10\") " pod="cert-manager/cert-manager-5b446d88c5-f9chx" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.124009 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8wvf\" (UniqueName: \"kubernetes.io/projected/1d540303-134e-4ca8-9597-9eb8dda29ef4-kube-api-access-m8wvf\") pod \"cert-manager-cainjector-7f985d654d-99ngt\" (UID: \"1d540303-134e-4ca8-9597-9eb8dda29ef4\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.124020 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcds4\" (UniqueName: \"kubernetes.io/projected/4fe2d6c0-b093-452f-8dfa-0db642716f10-kube-api-access-kcds4\") pod \"cert-manager-5b446d88c5-f9chx\" (UID: \"4fe2d6c0-b093-452f-8dfa-0db642716f10\") " pod="cert-manager/cert-manager-5b446d88c5-f9chx" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.131631 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48rk7\" (UniqueName: \"kubernetes.io/projected/8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd-kube-api-access-48rk7\") pod \"cert-manager-webhook-5655c58dd6-mn5gf\" (UID: \"8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.226959 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.279019 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-f9chx" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.289001 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.603217 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-f9chx"] Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.609819 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.668547 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-99ngt"] Nov 26 06:34:36 crc kubenswrapper[4775]: W1126 06:34:36.678187 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d540303_134e_4ca8_9597_9eb8dda29ef4.slice/crio-9978d20666dcade5b55ce7ebbac09f8de94e650168dce9acb0f3cded30c16040 WatchSource:0}: Error finding container 9978d20666dcade5b55ce7ebbac09f8de94e650168dce9acb0f3cded30c16040: Status 404 returned error can't find the container with id 9978d20666dcade5b55ce7ebbac09f8de94e650168dce9acb0f3cded30c16040 Nov 26 06:34:36 crc kubenswrapper[4775]: I1126 06:34:36.772999 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-mn5gf"] Nov 26 06:34:36 crc kubenswrapper[4775]: W1126 06:34:36.773065 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b00ec65_b72b_4b6c_ac2e_f10f9ddae6cd.slice/crio-1d0112b5d2481ef4fded5dabcc873f6aba614c31d280a87041a2334d549de527 WatchSource:0}: Error finding container 1d0112b5d2481ef4fded5dabcc873f6aba614c31d280a87041a2334d549de527: Status 404 returned error can't find the container with id 1d0112b5d2481ef4fded5dabcc873f6aba614c31d280a87041a2334d549de527 Nov 26 06:34:37 crc kubenswrapper[4775]: I1126 06:34:37.219157 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" event={"ID":"8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd","Type":"ContainerStarted","Data":"1d0112b5d2481ef4fded5dabcc873f6aba614c31d280a87041a2334d549de527"} Nov 26 06:34:37 crc kubenswrapper[4775]: I1126 06:34:37.222905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" event={"ID":"1d540303-134e-4ca8-9597-9eb8dda29ef4","Type":"ContainerStarted","Data":"9978d20666dcade5b55ce7ebbac09f8de94e650168dce9acb0f3cded30c16040"} Nov 26 06:34:37 crc kubenswrapper[4775]: I1126 06:34:37.228549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-f9chx" event={"ID":"4fe2d6c0-b093-452f-8dfa-0db642716f10","Type":"ContainerStarted","Data":"db64055e0735085d1b7f14b58809bd23f60be4a4e0755ddb0150d41cb681d7f9"} Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.245179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-f9chx" event={"ID":"4fe2d6c0-b093-452f-8dfa-0db642716f10","Type":"ContainerStarted","Data":"981255407cad50899c2ca841d116dedc0a6efacc0b7d64609bcbd5e6a166a1bf"} Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.246880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" event={"ID":"8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd","Type":"ContainerStarted","Data":"c7c819acffa3931e56409b6882e7cd7a97e1d12179e41164cfdec0fd731839c6"} Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.247203 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.249668 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" event={"ID":"1d540303-134e-4ca8-9597-9eb8dda29ef4","Type":"ContainerStarted","Data":"ce39dda2b8e050d5572a0863875a544246abd7227d7cdea9531f029660335347"} Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.263348 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-f9chx" podStartSLOduration=2.124818733 podStartE2EDuration="5.263323716s" podCreationTimestamp="2025-11-26 06:34:35 +0000 UTC" firstStartedPulling="2025-11-26 06:34:36.609571664 +0000 UTC m=+559.970875616" lastFinishedPulling="2025-11-26 06:34:39.748076637 +0000 UTC m=+563.109380599" observedRunningTime="2025-11-26 06:34:40.257400736 +0000 UTC m=+563.618704698" watchObservedRunningTime="2025-11-26 06:34:40.263323716 +0000 UTC m=+563.624627698" Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.278352 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" podStartSLOduration=2.30633685 podStartE2EDuration="5.278330641s" podCreationTimestamp="2025-11-26 06:34:35 +0000 UTC" firstStartedPulling="2025-11-26 06:34:36.775206132 +0000 UTC m=+560.136510094" lastFinishedPulling="2025-11-26 06:34:39.747199913 +0000 UTC m=+563.108503885" observedRunningTime="2025-11-26 06:34:40.276391858 +0000 UTC m=+563.637695810" watchObservedRunningTime="2025-11-26 06:34:40.278330641 +0000 UTC m=+563.639634613" Nov 26 06:34:40 crc kubenswrapper[4775]: I1126 06:34:40.312065 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-99ngt" podStartSLOduration=2.264399119 podStartE2EDuration="5.31203225s" podCreationTimestamp="2025-11-26 06:34:35 +0000 UTC" firstStartedPulling="2025-11-26 06:34:36.680871588 +0000 UTC m=+560.042175540" lastFinishedPulling="2025-11-26 06:34:39.728504709 +0000 UTC m=+563.089808671" observedRunningTime="2025-11-26 06:34:40.30868855 +0000 UTC m=+563.669992502" watchObservedRunningTime="2025-11-26 06:34:40.31203225 +0000 UTC m=+563.673336242" Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.292035 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-mn5gf" Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.661357 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f9lb8"] Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.661908 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-controller" containerID="cri-o://19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.662205 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="northd" containerID="cri-o://87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.662417 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-acl-logging" containerID="cri-o://3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.662317 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.662319 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="sbdb" containerID="cri-o://ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.662420 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-node" containerID="cri-o://32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.662385 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="nbdb" containerID="cri-o://de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" gracePeriod=30 Nov 26 06:34:46 crc kubenswrapper[4775]: I1126 06:34:46.712809 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" containerID="cri-o://e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" gracePeriod=30 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.003744 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/3.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.006647 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovn-acl-logging/0.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.007463 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovn-controller/0.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.008044 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.081817 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cgnfx"] Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.092960 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="sbdb" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093033 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="sbdb" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093081 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093100 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093123 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kubecfg-setup" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093143 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kubecfg-setup" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093173 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093192 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093212 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093231 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093283 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="northd" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093300 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="northd" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093335 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093352 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093389 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093405 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093446 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-node" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093463 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-node" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093486 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-acl-logging" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093503 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-acl-logging" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.093540 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="nbdb" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.093557 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="nbdb" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094153 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="northd" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094195 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094219 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094239 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094271 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094295 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovn-acl-logging" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094328 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094365 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094395 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="kube-rbac-proxy-node" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094431 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="nbdb" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.094462 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="sbdb" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.095041 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.095074 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.095116 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.095133 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.095640 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerName="ovnkube-controller" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.101134 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157035 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-script-lib\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157167 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-slash\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157236 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4g4l\" (UniqueName: \"kubernetes.io/projected/53d5d732-edab-4a9d-802c-4d9f0a2197ac-kube-api-access-c4g4l\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157276 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-systemd-units\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157350 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovn-node-metrics-cert\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157365 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-slash" (OuterVolumeSpecName: "host-slash") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157789 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-config\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157837 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-kubelet\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157858 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-netns\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157914 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-ovn\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157941 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-var-lib-openvswitch\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157969 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-bin\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157969 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157970 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157986 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.157992 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-log-socket\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158017 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-log-socket" (OuterVolumeSpecName: "log-socket") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158002 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158018 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158097 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-node-log\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158152 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-netd\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158142 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-node-log" (OuterVolumeSpecName: "node-log") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158184 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158187 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-systemd\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158238 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-openvswitch\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158270 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-ovn-kubernetes\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158296 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-env-overrides\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158329 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-etc-openvswitch\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158355 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-var-lib-cni-networks-ovn-kubernetes\") pod \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\" (UID: \"53d5d732-edab-4a9d-802c-4d9f0a2197ac\") " Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158379 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158410 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158436 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158526 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158526 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158573 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-etc-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-env-overrides\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158645 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j8jg\" (UniqueName: \"kubernetes.io/projected/5c949403-100e-4167-845d-802ee60663c0-kube-api-access-4j8jg\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-node-log\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158692 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-cni-netd\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158733 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158798 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-cni-bin\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158908 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-systemd\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158949 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-ovnkube-script-lib\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158951 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.158996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c949403-100e-4167-845d-802ee60663c0-ovn-node-metrics-cert\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-ovn\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159316 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-var-lib-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-log-socket\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159430 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-kubelet\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159492 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-systemd-units\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159542 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-run-netns\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159602 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-ovnkube-config\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-run-ovn-kubernetes\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-slash\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159934 4775 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.159976 4775 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-log-socket\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160004 4775 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-node-log\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160027 4775 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160050 4775 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160075 4775 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160099 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160116 4775 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160136 4775 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160153 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160171 4775 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-slash\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160188 4775 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160204 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160220 4775 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160236 4775 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160252 4775 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.160268 4775 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.164834 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d5d732-edab-4a9d-802c-4d9f0a2197ac-kube-api-access-c4g4l" (OuterVolumeSpecName: "kube-api-access-c4g4l") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "kube-api-access-c4g4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.164848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.175930 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "53d5d732-edab-4a9d-802c-4d9f0a2197ac" (UID: "53d5d732-edab-4a9d-802c-4d9f0a2197ac"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.261913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-env-overrides\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j8jg\" (UniqueName: \"kubernetes.io/projected/5c949403-100e-4167-845d-802ee60663c0-kube-api-access-4j8jg\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262064 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-node-log\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-cni-netd\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262154 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262196 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-cni-bin\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262256 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-systemd\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-ovnkube-script-lib\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262346 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c949403-100e-4167-845d-802ee60663c0-ovn-node-metrics-cert\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262388 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-ovn\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262438 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-log-socket\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-var-lib-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262533 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-kubelet\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262586 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-systemd-units\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262626 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-run-netns\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262686 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-ovnkube-config\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-run-ovn-kubernetes\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262850 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-slash\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.262978 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-etc-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263082 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4g4l\" (UniqueName: \"kubernetes.io/projected/53d5d732-edab-4a9d-802c-4d9f0a2197ac-kube-api-access-c4g4l\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263113 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/53d5d732-edab-4a9d-802c-4d9f0a2197ac-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263142 4775 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/53d5d732-edab-4a9d-802c-4d9f0a2197ac-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263212 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-etc-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-node-log\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263871 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-cni-netd\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263933 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.263991 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-cni-bin\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.264053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-systemd\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.264144 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-env-overrides\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.264441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-systemd-units\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.265282 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-ovnkube-script-lib\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.265387 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-run-netns\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.265663 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-slash\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.265839 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-run-ovn-kubernetes\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.265939 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.266014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-log-socket\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.266060 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-run-ovn\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.266108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-var-lib-openvswitch\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.266149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5c949403-100e-4167-845d-802ee60663c0-host-kubelet\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.266491 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c949403-100e-4167-845d-802ee60663c0-ovnkube-config\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.270823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c949403-100e-4167-845d-802ee60663c0-ovn-node-metrics-cert\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.293313 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j8jg\" (UniqueName: \"kubernetes.io/projected/5c949403-100e-4167-845d-802ee60663c0-kube-api-access-4j8jg\") pod \"ovnkube-node-cgnfx\" (UID: \"5c949403-100e-4167-845d-802ee60663c0\") " pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.302451 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/2.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.304744 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/1.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.304825 4775 generic.go:334] "Generic (PLEG): container finished" podID="4181bfb8-478b-4b30-a2cb-744da1010aeb" containerID="8dd11dfd36979f712eecdab186571b6768bcbde711a197ba3cd3465503d292ee" exitCode=2 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.304913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerDied","Data":"8dd11dfd36979f712eecdab186571b6768bcbde711a197ba3cd3465503d292ee"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.304957 4775 scope.go:117] "RemoveContainer" containerID="a9ea780d3ca3e0aa3e8bb64596470056612b57d1cefb46f80e276dbaaf4d5f94" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.305556 4775 scope.go:117] "RemoveContainer" containerID="8dd11dfd36979f712eecdab186571b6768bcbde711a197ba3cd3465503d292ee" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.306030 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-k9pcc_openshift-multus(4181bfb8-478b-4b30-a2cb-744da1010aeb)\"" pod="openshift-multus/multus-k9pcc" podUID="4181bfb8-478b-4b30-a2cb-744da1010aeb" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.317261 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovnkube-controller/3.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.321919 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovn-acl-logging/0.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.323140 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f9lb8_53d5d732-edab-4a9d-802c-4d9f0a2197ac/ovn-controller/0.log" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324542 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" exitCode=0 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324584 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" exitCode=0 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324608 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" exitCode=0 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324629 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" exitCode=0 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324677 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" exitCode=0 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324694 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" exitCode=0 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324749 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" exitCode=143 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324769 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" exitCode=143 Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324807 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324856 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324886 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324956 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324978 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.324998 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325011 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325026 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325040 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325053 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325067 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325081 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325095 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325110 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325129 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325149 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325164 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325178 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325192 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325205 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325218 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325230 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325243 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325256 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325269 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325306 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325320 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325333 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325348 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325361 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325374 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325390 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325403 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325416 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325430 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" event={"ID":"53d5d732-edab-4a9d-802c-4d9f0a2197ac","Type":"ContainerDied","Data":"1a2e13650cc86235ac977839ba308136ea2049d89bc454f98bd3cf310e6af2b4"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325468 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325483 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325498 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325510 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325523 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325536 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325549 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325561 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325575 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325587 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.325991 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f9lb8" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.358138 4775 scope.go:117] "RemoveContainer" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.384979 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.390280 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f9lb8"] Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.396024 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f9lb8"] Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.404249 4775 scope.go:117] "RemoveContainer" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.423672 4775 scope.go:117] "RemoveContainer" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.434852 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.436120 4775 scope.go:117] "RemoveContainer" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.451023 4775 scope.go:117] "RemoveContainer" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.469297 4775 scope.go:117] "RemoveContainer" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.488750 4775 scope.go:117] "RemoveContainer" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.511446 4775 scope.go:117] "RemoveContainer" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.537000 4775 scope.go:117] "RemoveContainer" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.569863 4775 scope.go:117] "RemoveContainer" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.570468 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": container with ID starting with e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c not found: ID does not exist" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.570525 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} err="failed to get container status \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": rpc error: code = NotFound desc = could not find container \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": container with ID starting with e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.570559 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.571119 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": container with ID starting with c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00 not found: ID does not exist" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.571160 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} err="failed to get container status \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": rpc error: code = NotFound desc = could not find container \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": container with ID starting with c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.571184 4775 scope.go:117] "RemoveContainer" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.571508 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": container with ID starting with ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523 not found: ID does not exist" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.571564 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} err="failed to get container status \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": rpc error: code = NotFound desc = could not find container \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": container with ID starting with ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.571596 4775 scope.go:117] "RemoveContainer" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.572108 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": container with ID starting with de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f not found: ID does not exist" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.572323 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} err="failed to get container status \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": rpc error: code = NotFound desc = could not find container \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": container with ID starting with de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.572407 4775 scope.go:117] "RemoveContainer" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.573043 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": container with ID starting with 87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0 not found: ID does not exist" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.573155 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} err="failed to get container status \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": rpc error: code = NotFound desc = could not find container \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": container with ID starting with 87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.573190 4775 scope.go:117] "RemoveContainer" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.573580 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": container with ID starting with 89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6 not found: ID does not exist" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.573673 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} err="failed to get container status \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": rpc error: code = NotFound desc = could not find container \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": container with ID starting with 89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.573702 4775 scope.go:117] "RemoveContainer" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.574090 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": container with ID starting with 32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16 not found: ID does not exist" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.574132 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} err="failed to get container status \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": rpc error: code = NotFound desc = could not find container \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": container with ID starting with 32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.574169 4775 scope.go:117] "RemoveContainer" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.574655 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": container with ID starting with 3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07 not found: ID does not exist" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.574694 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} err="failed to get container status \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": rpc error: code = NotFound desc = could not find container \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": container with ID starting with 3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.574759 4775 scope.go:117] "RemoveContainer" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.575069 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": container with ID starting with 19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5 not found: ID does not exist" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.575127 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} err="failed to get container status \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": rpc error: code = NotFound desc = could not find container \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": container with ID starting with 19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.575153 4775 scope.go:117] "RemoveContainer" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" Nov 26 06:34:47 crc kubenswrapper[4775]: E1126 06:34:47.575435 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": container with ID starting with edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975 not found: ID does not exist" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.575475 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} err="failed to get container status \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": rpc error: code = NotFound desc = could not find container \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": container with ID starting with edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.575500 4775 scope.go:117] "RemoveContainer" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.575889 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} err="failed to get container status \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": rpc error: code = NotFound desc = could not find container \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": container with ID starting with e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.575981 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.576314 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} err="failed to get container status \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": rpc error: code = NotFound desc = could not find container \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": container with ID starting with c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.576348 4775 scope.go:117] "RemoveContainer" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.576654 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} err="failed to get container status \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": rpc error: code = NotFound desc = could not find container \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": container with ID starting with ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.576701 4775 scope.go:117] "RemoveContainer" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.577212 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} err="failed to get container status \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": rpc error: code = NotFound desc = could not find container \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": container with ID starting with de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.577305 4775 scope.go:117] "RemoveContainer" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.577765 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} err="failed to get container status \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": rpc error: code = NotFound desc = could not find container \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": container with ID starting with 87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.577810 4775 scope.go:117] "RemoveContainer" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.580239 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} err="failed to get container status \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": rpc error: code = NotFound desc = could not find container \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": container with ID starting with 89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.580283 4775 scope.go:117] "RemoveContainer" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.580653 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} err="failed to get container status \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": rpc error: code = NotFound desc = could not find container \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": container with ID starting with 32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.580691 4775 scope.go:117] "RemoveContainer" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.581112 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} err="failed to get container status \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": rpc error: code = NotFound desc = could not find container \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": container with ID starting with 3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.581152 4775 scope.go:117] "RemoveContainer" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.581457 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} err="failed to get container status \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": rpc error: code = NotFound desc = could not find container \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": container with ID starting with 19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.581499 4775 scope.go:117] "RemoveContainer" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.581831 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} err="failed to get container status \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": rpc error: code = NotFound desc = could not find container \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": container with ID starting with edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.581873 4775 scope.go:117] "RemoveContainer" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.582160 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} err="failed to get container status \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": rpc error: code = NotFound desc = could not find container \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": container with ID starting with e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.582201 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.582527 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} err="failed to get container status \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": rpc error: code = NotFound desc = could not find container \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": container with ID starting with c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.582565 4775 scope.go:117] "RemoveContainer" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.582858 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} err="failed to get container status \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": rpc error: code = NotFound desc = could not find container \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": container with ID starting with ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.582896 4775 scope.go:117] "RemoveContainer" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.583203 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} err="failed to get container status \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": rpc error: code = NotFound desc = could not find container \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": container with ID starting with de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.583242 4775 scope.go:117] "RemoveContainer" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.583518 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} err="failed to get container status \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": rpc error: code = NotFound desc = could not find container \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": container with ID starting with 87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.583561 4775 scope.go:117] "RemoveContainer" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.584056 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} err="failed to get container status \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": rpc error: code = NotFound desc = could not find container \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": container with ID starting with 89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.584104 4775 scope.go:117] "RemoveContainer" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.584544 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} err="failed to get container status \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": rpc error: code = NotFound desc = could not find container \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": container with ID starting with 32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.584584 4775 scope.go:117] "RemoveContainer" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.584929 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} err="failed to get container status \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": rpc error: code = NotFound desc = could not find container \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": container with ID starting with 3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.584980 4775 scope.go:117] "RemoveContainer" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.585816 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} err="failed to get container status \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": rpc error: code = NotFound desc = could not find container \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": container with ID starting with 19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.585872 4775 scope.go:117] "RemoveContainer" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.586377 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} err="failed to get container status \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": rpc error: code = NotFound desc = could not find container \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": container with ID starting with edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.586415 4775 scope.go:117] "RemoveContainer" containerID="e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.586934 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c"} err="failed to get container status \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": rpc error: code = NotFound desc = could not find container \"e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c\": container with ID starting with e1e83ec9e59998e4f0c1af9763b19cbdcb27e479d9199cd00f82ba7e0fef9e5c not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.586973 4775 scope.go:117] "RemoveContainer" containerID="c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.587262 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00"} err="failed to get container status \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": rpc error: code = NotFound desc = could not find container \"c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00\": container with ID starting with c606cdccf77124009429e6c1678ca60023a77e8c1c4df7f73f41f5d18b2f5b00 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.587304 4775 scope.go:117] "RemoveContainer" containerID="ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.587576 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523"} err="failed to get container status \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": rpc error: code = NotFound desc = could not find container \"ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523\": container with ID starting with ae48a30624b214ca3f5d1972bbb2c696da2e6b8a774e97536bdd6fb48ed81523 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.587614 4775 scope.go:117] "RemoveContainer" containerID="de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.587967 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f"} err="failed to get container status \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": rpc error: code = NotFound desc = could not find container \"de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f\": container with ID starting with de93d8f1d649127a35af9c15c323f76916c97e57ec103c32dea15d0a6fbab55f not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.588003 4775 scope.go:117] "RemoveContainer" containerID="87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.588453 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0"} err="failed to get container status \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": rpc error: code = NotFound desc = could not find container \"87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0\": container with ID starting with 87813194b306c2a72ae20e95c85733f73cf80fee9160c0d9a066cf45f9eba6f0 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.588537 4775 scope.go:117] "RemoveContainer" containerID="89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.589175 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6"} err="failed to get container status \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": rpc error: code = NotFound desc = could not find container \"89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6\": container with ID starting with 89f977c3350f7881c95531d500aad63b27add721f253eea216fb8c00252329f6 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.589267 4775 scope.go:117] "RemoveContainer" containerID="32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.590125 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16"} err="failed to get container status \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": rpc error: code = NotFound desc = could not find container \"32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16\": container with ID starting with 32cbd07f6e6006d8e12ad48fd64dc36573360cf1c19d51dcb3e27e97efba0f16 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.590217 4775 scope.go:117] "RemoveContainer" containerID="3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.590789 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07"} err="failed to get container status \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": rpc error: code = NotFound desc = could not find container \"3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07\": container with ID starting with 3485af2123d601c93f89bbd3470571d7561654ba41d309a53461fced9ef23a07 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.590833 4775 scope.go:117] "RemoveContainer" containerID="19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.591283 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5"} err="failed to get container status \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": rpc error: code = NotFound desc = could not find container \"19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5\": container with ID starting with 19fe2989248807924bbf243317c021c03a1f14ad4e19b71aa67ab29d6c77d5a5 not found: ID does not exist" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.591320 4775 scope.go:117] "RemoveContainer" containerID="edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975" Nov 26 06:34:47 crc kubenswrapper[4775]: I1126 06:34:47.592815 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975"} err="failed to get container status \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": rpc error: code = NotFound desc = could not find container \"edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975\": container with ID starting with edc53b8f3603777e7ba4c696594ced9c69eb8d546491db96dcbcd5d66b280975 not found: ID does not exist" Nov 26 06:34:48 crc kubenswrapper[4775]: I1126 06:34:48.333950 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/2.log" Nov 26 06:34:48 crc kubenswrapper[4775]: I1126 06:34:48.338024 4775 generic.go:334] "Generic (PLEG): container finished" podID="5c949403-100e-4167-845d-802ee60663c0" containerID="073ee38e7b7eaa3e3c4c073e27394199c2d1ef59d200a28cc069389861dda71f" exitCode=0 Nov 26 06:34:48 crc kubenswrapper[4775]: I1126 06:34:48.338098 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerDied","Data":"073ee38e7b7eaa3e3c4c073e27394199c2d1ef59d200a28cc069389861dda71f"} Nov 26 06:34:48 crc kubenswrapper[4775]: I1126 06:34:48.338182 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"15b2ae49a120c8c146b5dd33c8fb088acae3aa6c2add5d96613fee9538a5a02d"} Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.337545 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d5d732-edab-4a9d-802c-4d9f0a2197ac" path="/var/lib/kubelet/pods/53d5d732-edab-4a9d-802c-4d9f0a2197ac/volumes" Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.354227 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"8081871da36e8a2b8661403661e9a2ebb5974291d709c2767a24a5464dd9bad6"} Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.354279 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"0f6482c45835cf66915eda64d9712ef0d07b6e90ef419de33b619dbb936f45fa"} Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.354297 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"54be0e51b076960db12fe5b16aa302ab20f5311de838020b540044a180cbe41d"} Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.354311 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"9dc8c1ecfde7fda20369b0a570ae49bbb4988a1e6732eec752d3f4e7a45111fa"} Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.354323 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"d33479a90585835d5750b6452e52463dec6f2b5bd260e85c2e8957a98c365c02"} Nov 26 06:34:49 crc kubenswrapper[4775]: I1126 06:34:49.354335 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"1d3f9ceac7b0ffc985ac9d390c7fc90f29f8a0bc9b94c0c2097c97194a51e654"} Nov 26 06:34:51 crc kubenswrapper[4775]: I1126 06:34:51.419980 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:34:51 crc kubenswrapper[4775]: I1126 06:34:51.420103 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:34:51 crc kubenswrapper[4775]: I1126 06:34:51.420189 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:34:51 crc kubenswrapper[4775]: I1126 06:34:51.421518 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"988eb2b1e284ef66703f1f37b2da654fd669430dce2be08f52d009293283f65e"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:34:51 crc kubenswrapper[4775]: I1126 06:34:51.421711 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://988eb2b1e284ef66703f1f37b2da654fd669430dce2be08f52d009293283f65e" gracePeriod=600 Nov 26 06:34:52 crc kubenswrapper[4775]: I1126 06:34:52.381656 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"b9e3a570edd32ab602e818c26a83d95caebded41de98d440f06898ddd88059e7"} Nov 26 06:34:52 crc kubenswrapper[4775]: I1126 06:34:52.386101 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="988eb2b1e284ef66703f1f37b2da654fd669430dce2be08f52d009293283f65e" exitCode=0 Nov 26 06:34:52 crc kubenswrapper[4775]: I1126 06:34:52.386166 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"988eb2b1e284ef66703f1f37b2da654fd669430dce2be08f52d009293283f65e"} Nov 26 06:34:52 crc kubenswrapper[4775]: I1126 06:34:52.386245 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"8685d35e684f235bf0161f8f3e1dacf8484e59a0d24a51ad51c239999335747e"} Nov 26 06:34:52 crc kubenswrapper[4775]: I1126 06:34:52.386271 4775 scope.go:117] "RemoveContainer" containerID="a22b92d95ea00f90fd226634f524ff08e52cb46e50050a0f9a3125f6b85b48ba" Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.411435 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" event={"ID":"5c949403-100e-4167-845d-802ee60663c0","Type":"ContainerStarted","Data":"22619b55d51ed9bf79d5e109acebd4b3700c33bc858f1b597c02a85f5ae58365"} Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.412397 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.412531 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.412663 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.456402 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" podStartSLOduration=7.456379172 podStartE2EDuration="7.456379172s" podCreationTimestamp="2025-11-26 06:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:34:54.453665698 +0000 UTC m=+577.814969700" watchObservedRunningTime="2025-11-26 06:34:54.456379172 +0000 UTC m=+577.817683164" Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.464225 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:34:54 crc kubenswrapper[4775]: I1126 06:34:54.468346 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:35:02 crc kubenswrapper[4775]: I1126 06:35:02.328284 4775 scope.go:117] "RemoveContainer" containerID="8dd11dfd36979f712eecdab186571b6768bcbde711a197ba3cd3465503d292ee" Nov 26 06:35:02 crc kubenswrapper[4775]: E1126 06:35:02.328937 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-k9pcc_openshift-multus(4181bfb8-478b-4b30-a2cb-744da1010aeb)\"" pod="openshift-multus/multus-k9pcc" podUID="4181bfb8-478b-4b30-a2cb-744da1010aeb" Nov 26 06:35:13 crc kubenswrapper[4775]: I1126 06:35:13.327906 4775 scope.go:117] "RemoveContainer" containerID="8dd11dfd36979f712eecdab186571b6768bcbde711a197ba3cd3465503d292ee" Nov 26 06:35:13 crc kubenswrapper[4775]: I1126 06:35:13.555909 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k9pcc_4181bfb8-478b-4b30-a2cb-744da1010aeb/kube-multus/2.log" Nov 26 06:35:13 crc kubenswrapper[4775]: I1126 06:35:13.556559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k9pcc" event={"ID":"4181bfb8-478b-4b30-a2cb-744da1010aeb","Type":"ContainerStarted","Data":"3e77fdf2dca30071f731da035bcfcc3410d1efa17ff14b58e0c7798cc0022824"} Nov 26 06:35:17 crc kubenswrapper[4775]: I1126 06:35:17.472454 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cgnfx" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.436991 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57"] Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.438682 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.445095 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.447085 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57"] Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.541902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.542002 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwsqv\" (UniqueName: \"kubernetes.io/projected/4ba6f052-9efe-4347-9394-922477db3e0e-kube-api-access-dwsqv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.542153 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.643746 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.644004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwsqv\" (UniqueName: \"kubernetes.io/projected/4ba6f052-9efe-4347-9394-922477db3e0e-kube-api-access-dwsqv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.644128 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.644607 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.644827 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.665206 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwsqv\" (UniqueName: \"kubernetes.io/projected/4ba6f052-9efe-4347-9394-922477db3e0e-kube-api-access-dwsqv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:25 crc kubenswrapper[4775]: I1126 06:35:25.789976 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:26 crc kubenswrapper[4775]: I1126 06:35:26.033129 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57"] Nov 26 06:35:26 crc kubenswrapper[4775]: I1126 06:35:26.640981 4775 generic.go:334] "Generic (PLEG): container finished" podID="4ba6f052-9efe-4347-9394-922477db3e0e" containerID="519e2cbf2ac91e6fdded939f6c8dbb9c3b1c9cbf4a399fab777f0f1198ec3a3b" exitCode=0 Nov 26 06:35:26 crc kubenswrapper[4775]: I1126 06:35:26.641049 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" event={"ID":"4ba6f052-9efe-4347-9394-922477db3e0e","Type":"ContainerDied","Data":"519e2cbf2ac91e6fdded939f6c8dbb9c3b1c9cbf4a399fab777f0f1198ec3a3b"} Nov 26 06:35:26 crc kubenswrapper[4775]: I1126 06:35:26.641091 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" event={"ID":"4ba6f052-9efe-4347-9394-922477db3e0e","Type":"ContainerStarted","Data":"b6389c6743048e540774ef84fb67a62c44e9290f8ebcb6e77916a67e6f89d4b7"} Nov 26 06:35:28 crc kubenswrapper[4775]: I1126 06:35:28.658239 4775 generic.go:334] "Generic (PLEG): container finished" podID="4ba6f052-9efe-4347-9394-922477db3e0e" containerID="b17ec77845dbfa9bc187e5a1e648a87416f43ee05ddf365c6ed84371c4a864ce" exitCode=0 Nov 26 06:35:28 crc kubenswrapper[4775]: I1126 06:35:28.658309 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" event={"ID":"4ba6f052-9efe-4347-9394-922477db3e0e","Type":"ContainerDied","Data":"b17ec77845dbfa9bc187e5a1e648a87416f43ee05ddf365c6ed84371c4a864ce"} Nov 26 06:35:29 crc kubenswrapper[4775]: I1126 06:35:29.667829 4775 generic.go:334] "Generic (PLEG): container finished" podID="4ba6f052-9efe-4347-9394-922477db3e0e" containerID="acef6da16272bd5d43f58d591011adc0cc207d36e1d3387aea010e48073354c8" exitCode=0 Nov 26 06:35:29 crc kubenswrapper[4775]: I1126 06:35:29.667961 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" event={"ID":"4ba6f052-9efe-4347-9394-922477db3e0e","Type":"ContainerDied","Data":"acef6da16272bd5d43f58d591011adc0cc207d36e1d3387aea010e48073354c8"} Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.008494 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.124404 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-util\") pod \"4ba6f052-9efe-4347-9394-922477db3e0e\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.124538 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwsqv\" (UniqueName: \"kubernetes.io/projected/4ba6f052-9efe-4347-9394-922477db3e0e-kube-api-access-dwsqv\") pod \"4ba6f052-9efe-4347-9394-922477db3e0e\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.124589 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-bundle\") pod \"4ba6f052-9efe-4347-9394-922477db3e0e\" (UID: \"4ba6f052-9efe-4347-9394-922477db3e0e\") " Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.125527 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-bundle" (OuterVolumeSpecName: "bundle") pod "4ba6f052-9efe-4347-9394-922477db3e0e" (UID: "4ba6f052-9efe-4347-9394-922477db3e0e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.132601 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba6f052-9efe-4347-9394-922477db3e0e-kube-api-access-dwsqv" (OuterVolumeSpecName: "kube-api-access-dwsqv") pod "4ba6f052-9efe-4347-9394-922477db3e0e" (UID: "4ba6f052-9efe-4347-9394-922477db3e0e"). InnerVolumeSpecName "kube-api-access-dwsqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.153953 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-util" (OuterVolumeSpecName: "util") pod "4ba6f052-9efe-4347-9394-922477db3e0e" (UID: "4ba6f052-9efe-4347-9394-922477db3e0e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.226996 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwsqv\" (UniqueName: \"kubernetes.io/projected/4ba6f052-9efe-4347-9394-922477db3e0e-kube-api-access-dwsqv\") on node \"crc\" DevicePath \"\"" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.227067 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.227093 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ba6f052-9efe-4347-9394-922477db3e0e-util\") on node \"crc\" DevicePath \"\"" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.685498 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" event={"ID":"4ba6f052-9efe-4347-9394-922477db3e0e","Type":"ContainerDied","Data":"b6389c6743048e540774ef84fb67a62c44e9290f8ebcb6e77916a67e6f89d4b7"} Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.685867 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6389c6743048e540774ef84fb67a62c44e9290f8ebcb6e77916a67e6f89d4b7" Nov 26 06:35:31 crc kubenswrapper[4775]: I1126 06:35:31.685593 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.953762 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-jhrbn"] Nov 26 06:35:32 crc kubenswrapper[4775]: E1126 06:35:32.953962 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="util" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.953973 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="util" Nov 26 06:35:32 crc kubenswrapper[4775]: E1126 06:35:32.953986 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="extract" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.953992 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="extract" Nov 26 06:35:32 crc kubenswrapper[4775]: E1126 06:35:32.954004 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="pull" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.954010 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="pull" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.954105 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba6f052-9efe-4347-9394-922477db3e0e" containerName="extract" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.954442 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.956898 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-d7thq" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.957044 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.959737 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 26 06:35:32 crc kubenswrapper[4775]: I1126 06:35:32.963799 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-jhrbn"] Nov 26 06:35:33 crc kubenswrapper[4775]: I1126 06:35:33.152754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9b7n\" (UniqueName: \"kubernetes.io/projected/5b010a92-1985-467a-9d70-35d65f48f13d-kube-api-access-t9b7n\") pod \"nmstate-operator-557fdffb88-jhrbn\" (UID: \"5b010a92-1985-467a-9d70-35d65f48f13d\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" Nov 26 06:35:33 crc kubenswrapper[4775]: I1126 06:35:33.254825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9b7n\" (UniqueName: \"kubernetes.io/projected/5b010a92-1985-467a-9d70-35d65f48f13d-kube-api-access-t9b7n\") pod \"nmstate-operator-557fdffb88-jhrbn\" (UID: \"5b010a92-1985-467a-9d70-35d65f48f13d\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" Nov 26 06:35:33 crc kubenswrapper[4775]: I1126 06:35:33.272796 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9b7n\" (UniqueName: \"kubernetes.io/projected/5b010a92-1985-467a-9d70-35d65f48f13d-kube-api-access-t9b7n\") pod \"nmstate-operator-557fdffb88-jhrbn\" (UID: \"5b010a92-1985-467a-9d70-35d65f48f13d\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" Nov 26 06:35:33 crc kubenswrapper[4775]: I1126 06:35:33.314192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" Nov 26 06:35:33 crc kubenswrapper[4775]: I1126 06:35:33.497042 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-jhrbn"] Nov 26 06:35:33 crc kubenswrapper[4775]: W1126 06:35:33.500004 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b010a92_1985_467a_9d70_35d65f48f13d.slice/crio-34871f2af6a812cd853aec13d15447bcc328827c1a057fab9225745bc2d55648 WatchSource:0}: Error finding container 34871f2af6a812cd853aec13d15447bcc328827c1a057fab9225745bc2d55648: Status 404 returned error can't find the container with id 34871f2af6a812cd853aec13d15447bcc328827c1a057fab9225745bc2d55648 Nov 26 06:35:33 crc kubenswrapper[4775]: I1126 06:35:33.700110 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" event={"ID":"5b010a92-1985-467a-9d70-35d65f48f13d","Type":"ContainerStarted","Data":"34871f2af6a812cd853aec13d15447bcc328827c1a057fab9225745bc2d55648"} Nov 26 06:35:35 crc kubenswrapper[4775]: I1126 06:35:35.712813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" event={"ID":"5b010a92-1985-467a-9d70-35d65f48f13d","Type":"ContainerStarted","Data":"12e85b7c25e07710b3d917acc09c90e9025e5f227fef64c8114225e22a925d6a"} Nov 26 06:35:35 crc kubenswrapper[4775]: I1126 06:35:35.727466 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-jhrbn" podStartSLOduration=1.706410933 podStartE2EDuration="3.727447095s" podCreationTimestamp="2025-11-26 06:35:32 +0000 UTC" firstStartedPulling="2025-11-26 06:35:33.501985881 +0000 UTC m=+616.863289833" lastFinishedPulling="2025-11-26 06:35:35.523022043 +0000 UTC m=+618.884325995" observedRunningTime="2025-11-26 06:35:35.7255803 +0000 UTC m=+619.086884242" watchObservedRunningTime="2025-11-26 06:35:35.727447095 +0000 UTC m=+619.088751047" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.666203 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.667467 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.670452 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-h6fgd" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.679534 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.680555 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.682252 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.698706 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2f477e8a-59b8-460b-b84d-6b1a4fd50827-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.698927 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw5gr\" (UniqueName: \"kubernetes.io/projected/910b3c41-e6e8-4c07-945d-d9c580ccce6c-kube-api-access-vw5gr\") pod \"nmstate-metrics-5dcf9c57c5-2xn6f\" (UID: \"910b3c41-e6e8-4c07-945d-d9c580ccce6c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.699018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wh8w\" (UniqueName: \"kubernetes.io/projected/2f477e8a-59b8-460b-b84d-6b1a4fd50827-kube-api-access-8wh8w\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.707046 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.713384 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-2tzgw"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.714419 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.752008 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.800444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wh8w\" (UniqueName: \"kubernetes.io/projected/2f477e8a-59b8-460b-b84d-6b1a4fd50827-kube-api-access-8wh8w\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.800830 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2f477e8a-59b8-460b-b84d-6b1a4fd50827-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.801063 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw5gr\" (UniqueName: \"kubernetes.io/projected/910b3c41-e6e8-4c07-945d-d9c580ccce6c-kube-api-access-vw5gr\") pod \"nmstate-metrics-5dcf9c57c5-2xn6f\" (UID: \"910b3c41-e6e8-4c07-945d-d9c580ccce6c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" Nov 26 06:35:36 crc kubenswrapper[4775]: E1126 06:35:36.800961 4775 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 26 06:35:36 crc kubenswrapper[4775]: E1126 06:35:36.801304 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f477e8a-59b8-460b-b84d-6b1a4fd50827-tls-key-pair podName:2f477e8a-59b8-460b-b84d-6b1a4fd50827 nodeName:}" failed. No retries permitted until 2025-11-26 06:35:37.301275495 +0000 UTC m=+620.662579437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/2f477e8a-59b8-460b-b84d-6b1a4fd50827-tls-key-pair") pod "nmstate-webhook-6b89b748d8-c49w5" (UID: "2f477e8a-59b8-460b-b84d-6b1a4fd50827") : secret "openshift-nmstate-webhook" not found Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.825737 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw5gr\" (UniqueName: \"kubernetes.io/projected/910b3c41-e6e8-4c07-945d-d9c580ccce6c-kube-api-access-vw5gr\") pod \"nmstate-metrics-5dcf9c57c5-2xn6f\" (UID: \"910b3c41-e6e8-4c07-945d-d9c580ccce6c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.833222 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wh8w\" (UniqueName: \"kubernetes.io/projected/2f477e8a-59b8-460b-b84d-6b1a4fd50827-kube-api-access-8wh8w\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.850390 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.851769 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.868805 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vw4bd" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.869009 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.870457 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.880641 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr"] Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.902416 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-dbus-socket\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.902560 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79qv2\" (UniqueName: \"kubernetes.io/projected/98657427-74ec-48e6-89d9-ef27646c475a-kube-api-access-79qv2\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.902612 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-ovs-socket\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.902701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-nmstate-lock\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:36 crc kubenswrapper[4775]: I1126 06:35:36.989590 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004511 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79qv2\" (UniqueName: \"kubernetes.io/projected/98657427-74ec-48e6-89d9-ef27646c475a-kube-api-access-79qv2\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-ovs-socket\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004613 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zzp8\" (UniqueName: \"kubernetes.io/projected/d1e0c343-feef-4493-bec5-d725f665b9d6-kube-api-access-5zzp8\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1e0c343-feef-4493-bec5-d725f665b9d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004665 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-nmstate-lock\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004728 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-ovs-socket\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004783 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d1e0c343-feef-4493-bec5-d725f665b9d6-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004825 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-nmstate-lock\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.004889 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-dbus-socket\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.005108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/98657427-74ec-48e6-89d9-ef27646c475a-dbus-socket\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.019421 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5dddd7d494-5g898"] Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.020248 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.039328 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dddd7d494-5g898"] Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.052578 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79qv2\" (UniqueName: \"kubernetes.io/projected/98657427-74ec-48e6-89d9-ef27646c475a-kube-api-access-79qv2\") pod \"nmstate-handler-2tzgw\" (UID: \"98657427-74ec-48e6-89d9-ef27646c475a\") " pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.057033 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:37 crc kubenswrapper[4775]: W1126 06:35:37.097787 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98657427_74ec_48e6_89d9_ef27646c475a.slice/crio-2106a25134a33ab0a7722cfc01126b366095ee8c000dbb6b9b115e80faa03cf7 WatchSource:0}: Error finding container 2106a25134a33ab0a7722cfc01126b366095ee8c000dbb6b9b115e80faa03cf7: Status 404 returned error can't find the container with id 2106a25134a33ab0a7722cfc01126b366095ee8c000dbb6b9b115e80faa03cf7 Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1e0c343-feef-4493-bec5-d725f665b9d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106309 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zzp8\" (UniqueName: \"kubernetes.io/projected/d1e0c343-feef-4493-bec5-d725f665b9d6-kube-api-access-5zzp8\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106330 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-serving-cert\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-config\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106387 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmmvq\" (UniqueName: \"kubernetes.io/projected/0bb52306-86d1-4cfc-8b3b-bd6addd05687-kube-api-access-qmmvq\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106407 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-oauth-serving-cert\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106424 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-service-ca\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d1e0c343-feef-4493-bec5-d725f665b9d6-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106483 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-oauth-config\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.106501 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-trusted-ca-bundle\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.108270 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d1e0c343-feef-4493-bec5-d725f665b9d6-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.109413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d1e0c343-feef-4493-bec5-d725f665b9d6-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.134308 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zzp8\" (UniqueName: \"kubernetes.io/projected/d1e0c343-feef-4493-bec5-d725f665b9d6-kube-api-access-5zzp8\") pod \"nmstate-console-plugin-5874bd7bc5-k5dwr\" (UID: \"d1e0c343-feef-4493-bec5-d725f665b9d6\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.189067 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-oauth-config\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207452 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-trusted-ca-bundle\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207476 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-serving-cert\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-config\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207534 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmmvq\" (UniqueName: \"kubernetes.io/projected/0bb52306-86d1-4cfc-8b3b-bd6addd05687-kube-api-access-qmmvq\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207555 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-oauth-serving-cert\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.207572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-service-ca\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.208756 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-service-ca\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.208996 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-config\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.209366 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-trusted-ca-bundle\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.210303 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0bb52306-86d1-4cfc-8b3b-bd6addd05687-oauth-serving-cert\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.210501 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-oauth-config\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.211938 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0bb52306-86d1-4cfc-8b3b-bd6addd05687-console-serving-cert\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.220192 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f"] Nov 26 06:35:37 crc kubenswrapper[4775]: W1126 06:35:37.222255 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod910b3c41_e6e8_4c07_945d_d9c580ccce6c.slice/crio-4c9d4de24c9f7c53c5ab81d38f58d40edfac6d0476fa67cf2f4f9e9f0dd84fe8 WatchSource:0}: Error finding container 4c9d4de24c9f7c53c5ab81d38f58d40edfac6d0476fa67cf2f4f9e9f0dd84fe8: Status 404 returned error can't find the container with id 4c9d4de24c9f7c53c5ab81d38f58d40edfac6d0476fa67cf2f4f9e9f0dd84fe8 Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.226951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmmvq\" (UniqueName: \"kubernetes.io/projected/0bb52306-86d1-4cfc-8b3b-bd6addd05687-kube-api-access-qmmvq\") pod \"console-5dddd7d494-5g898\" (UID: \"0bb52306-86d1-4cfc-8b3b-bd6addd05687\") " pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.308091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2f477e8a-59b8-460b-b84d-6b1a4fd50827-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.313404 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2f477e8a-59b8-460b-b84d-6b1a4fd50827-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-c49w5\" (UID: \"2f477e8a-59b8-460b-b84d-6b1a4fd50827\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.348323 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.373554 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr"] Nov 26 06:35:37 crc kubenswrapper[4775]: W1126 06:35:37.378904 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1e0c343_feef_4493_bec5_d725f665b9d6.slice/crio-3fd12f71eb82e3291d449abfc8e2de689d54ff61ea1b3622613751c246db5f39 WatchSource:0}: Error finding container 3fd12f71eb82e3291d449abfc8e2de689d54ff61ea1b3622613751c246db5f39: Status 404 returned error can't find the container with id 3fd12f71eb82e3291d449abfc8e2de689d54ff61ea1b3622613751c246db5f39 Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.379782 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.601592 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5"] Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.634800 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dddd7d494-5g898"] Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.722922 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2tzgw" event={"ID":"98657427-74ec-48e6-89d9-ef27646c475a","Type":"ContainerStarted","Data":"2106a25134a33ab0a7722cfc01126b366095ee8c000dbb6b9b115e80faa03cf7"} Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.724314 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" event={"ID":"910b3c41-e6e8-4c07-945d-d9c580ccce6c","Type":"ContainerStarted","Data":"4c9d4de24c9f7c53c5ab81d38f58d40edfac6d0476fa67cf2f4f9e9f0dd84fe8"} Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.726098 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" event={"ID":"2f477e8a-59b8-460b-b84d-6b1a4fd50827","Type":"ContainerStarted","Data":"2f165624553b779c5e38212909d98f23aa528f21073690bd7c5310fef6a987f9"} Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.726887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" event={"ID":"d1e0c343-feef-4493-bec5-d725f665b9d6","Type":"ContainerStarted","Data":"3fd12f71eb82e3291d449abfc8e2de689d54ff61ea1b3622613751c246db5f39"} Nov 26 06:35:37 crc kubenswrapper[4775]: I1126 06:35:37.727579 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dddd7d494-5g898" event={"ID":"0bb52306-86d1-4cfc-8b3b-bd6addd05687","Type":"ContainerStarted","Data":"39b37d697cc0a829a1086f7136bdeea134656f11daf0d3fceda2f20984b6100f"} Nov 26 06:35:38 crc kubenswrapper[4775]: I1126 06:35:38.736313 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dddd7d494-5g898" event={"ID":"0bb52306-86d1-4cfc-8b3b-bd6addd05687","Type":"ContainerStarted","Data":"9704e62cf6033c5641858159dd82c9360744b658f3d09c35e05cb79038c153bd"} Nov 26 06:35:38 crc kubenswrapper[4775]: I1126 06:35:38.751260 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5dddd7d494-5g898" podStartSLOduration=2.751244792 podStartE2EDuration="2.751244792s" podCreationTimestamp="2025-11-26 06:35:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:35:38.750011136 +0000 UTC m=+622.111315098" watchObservedRunningTime="2025-11-26 06:35:38.751244792 +0000 UTC m=+622.112548744" Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.751955 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2tzgw" event={"ID":"98657427-74ec-48e6-89d9-ef27646c475a","Type":"ContainerStarted","Data":"f8e7bf575f97899852c6d45036a05b57c773a2e1d94667c1f98dd2dfff466b0c"} Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.752547 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.755353 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" event={"ID":"910b3c41-e6e8-4c07-945d-d9c580ccce6c","Type":"ContainerStarted","Data":"d767d17cb606822281f12d8134546cd1a23a580ae85fb499e25e10fd3c707787"} Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.759106 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" event={"ID":"2f477e8a-59b8-460b-b84d-6b1a4fd50827","Type":"ContainerStarted","Data":"4f1f9cd01a5012aeb3635cec56266c0dad2d09ede5df52920b91c72147d39a43"} Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.759296 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.760260 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" event={"ID":"d1e0c343-feef-4493-bec5-d725f665b9d6","Type":"ContainerStarted","Data":"585260a9791cffd6ed643192785d61fbda0af5adeac4f9eaf5518d917e5783ad"} Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.773897 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-2tzgw" podStartSLOduration=1.9367087349999998 podStartE2EDuration="4.773877289s" podCreationTimestamp="2025-11-26 06:35:36 +0000 UTC" firstStartedPulling="2025-11-26 06:35:37.099826922 +0000 UTC m=+620.461130874" lastFinishedPulling="2025-11-26 06:35:39.936995486 +0000 UTC m=+623.298299428" observedRunningTime="2025-11-26 06:35:40.76768882 +0000 UTC m=+624.128992792" watchObservedRunningTime="2025-11-26 06:35:40.773877289 +0000 UTC m=+624.135181251" Nov 26 06:35:40 crc kubenswrapper[4775]: I1126 06:35:40.790197 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-k5dwr" podStartSLOduration=2.252644728 podStartE2EDuration="4.790176563s" podCreationTimestamp="2025-11-26 06:35:36 +0000 UTC" firstStartedPulling="2025-11-26 06:35:37.381487279 +0000 UTC m=+620.742791231" lastFinishedPulling="2025-11-26 06:35:39.919019104 +0000 UTC m=+623.280323066" observedRunningTime="2025-11-26 06:35:40.780204963 +0000 UTC m=+624.141508985" watchObservedRunningTime="2025-11-26 06:35:40.790176563 +0000 UTC m=+624.151480515" Nov 26 06:35:42 crc kubenswrapper[4775]: I1126 06:35:42.777933 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" event={"ID":"910b3c41-e6e8-4c07-945d-d9c580ccce6c","Type":"ContainerStarted","Data":"5652dea669e429d73f25a0ebbdcaf201e70eca07db4a7274af8b2cd467eacda9"} Nov 26 06:35:42 crc kubenswrapper[4775]: I1126 06:35:42.798119 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-2xn6f" podStartSLOduration=1.471821635 podStartE2EDuration="6.798103664s" podCreationTimestamp="2025-11-26 06:35:36 +0000 UTC" firstStartedPulling="2025-11-26 06:35:37.224037293 +0000 UTC m=+620.585341245" lastFinishedPulling="2025-11-26 06:35:42.550319322 +0000 UTC m=+625.911623274" observedRunningTime="2025-11-26 06:35:42.797501807 +0000 UTC m=+626.158805769" watchObservedRunningTime="2025-11-26 06:35:42.798103664 +0000 UTC m=+626.159407616" Nov 26 06:35:42 crc kubenswrapper[4775]: I1126 06:35:42.798570 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" podStartSLOduration=4.486008062 podStartE2EDuration="6.798565368s" podCreationTimestamp="2025-11-26 06:35:36 +0000 UTC" firstStartedPulling="2025-11-26 06:35:37.610807094 +0000 UTC m=+620.972111046" lastFinishedPulling="2025-11-26 06:35:39.92336437 +0000 UTC m=+623.284668352" observedRunningTime="2025-11-26 06:35:40.801162272 +0000 UTC m=+624.162466284" watchObservedRunningTime="2025-11-26 06:35:42.798565368 +0000 UTC m=+626.159869330" Nov 26 06:35:47 crc kubenswrapper[4775]: I1126 06:35:47.093894 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-2tzgw" Nov 26 06:35:47 crc kubenswrapper[4775]: I1126 06:35:47.380886 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:47 crc kubenswrapper[4775]: I1126 06:35:47.380967 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:47 crc kubenswrapper[4775]: I1126 06:35:47.386874 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:47 crc kubenswrapper[4775]: I1126 06:35:47.823146 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5dddd7d494-5g898" Nov 26 06:35:47 crc kubenswrapper[4775]: I1126 06:35:47.896688 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vl6j8"] Nov 26 06:35:57 crc kubenswrapper[4775]: I1126 06:35:57.357591 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-c49w5" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.615341 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z"] Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.617431 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.619819 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.634255 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z"] Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.734802 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.734907 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.734997 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvb82\" (UniqueName: \"kubernetes.io/projected/b4e29c14-21e8-4dc0-a701-493ae7fbda58-kube-api-access-cvb82\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.839320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvb82\" (UniqueName: \"kubernetes.io/projected/b4e29c14-21e8-4dc0-a701-493ae7fbda58-kube-api-access-cvb82\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.839612 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.839704 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.842877 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.843422 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.889551 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvb82\" (UniqueName: \"kubernetes.io/projected/b4e29c14-21e8-4dc0-a701-493ae7fbda58-kube-api-access-cvb82\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:11 crc kubenswrapper[4775]: I1126 06:36:11.955118 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:12 crc kubenswrapper[4775]: I1126 06:36:12.206472 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z"] Nov 26 06:36:12 crc kubenswrapper[4775]: I1126 06:36:12.942677 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vl6j8" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerName="console" containerID="cri-o://8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e" gracePeriod=15 Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.008852 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerID="cb5035db41aa97b247d8ef6e2822314db36275d6f0ea24b9d1ca18ff45975729" exitCode=0 Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.008893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" event={"ID":"b4e29c14-21e8-4dc0-a701-493ae7fbda58","Type":"ContainerDied","Data":"cb5035db41aa97b247d8ef6e2822314db36275d6f0ea24b9d1ca18ff45975729"} Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.008920 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" event={"ID":"b4e29c14-21e8-4dc0-a701-493ae7fbda58","Type":"ContainerStarted","Data":"6eb601b644fa7a600b970f7aa5ab79c4fe4606fc97043e41a32561aeb86eafb0"} Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.358248 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vl6j8_268d85ad-f165-4cd2-931f-1da2caefe3cd/console/0.log" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.358322 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.460640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-trusted-ca-bundle\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461161 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-oauth-config\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461264 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtbj2\" (UniqueName: \"kubernetes.io/projected/268d85ad-f165-4cd2-931f-1da2caefe3cd-kube-api-access-wtbj2\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-config\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461353 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-serving-cert\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461416 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-oauth-serving-cert\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461449 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-service-ca\") pod \"268d85ad-f165-4cd2-931f-1da2caefe3cd\" (UID: \"268d85ad-f165-4cd2-931f-1da2caefe3cd\") " Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.461997 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.462247 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-service-ca" (OuterVolumeSpecName: "service-ca") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.462409 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-config" (OuterVolumeSpecName: "console-config") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.462568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.468623 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/268d85ad-f165-4cd2-931f-1da2caefe3cd-kube-api-access-wtbj2" (OuterVolumeSpecName: "kube-api-access-wtbj2") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "kube-api-access-wtbj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.468945 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.474856 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "268d85ad-f165-4cd2-931f-1da2caefe3cd" (UID: "268d85ad-f165-4cd2-931f-1da2caefe3cd"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563120 4775 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563164 4775 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563177 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563189 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563201 4775 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563212 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtbj2\" (UniqueName: \"kubernetes.io/projected/268d85ad-f165-4cd2-931f-1da2caefe3cd-kube-api-access-wtbj2\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:13 crc kubenswrapper[4775]: I1126 06:36:13.563229 4775 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/268d85ad-f165-4cd2-931f-1da2caefe3cd-console-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.019351 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vl6j8_268d85ad-f165-4cd2-931f-1da2caefe3cd/console/0.log" Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.019424 4775 generic.go:334] "Generic (PLEG): container finished" podID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerID="8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e" exitCode=2 Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.019469 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vl6j8" event={"ID":"268d85ad-f165-4cd2-931f-1da2caefe3cd","Type":"ContainerDied","Data":"8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e"} Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.019507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vl6j8" event={"ID":"268d85ad-f165-4cd2-931f-1da2caefe3cd","Type":"ContainerDied","Data":"c6e390cd4c627d8b833dd438f50c7494f534145c2887be6dbcd9b0f176c89724"} Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.019535 4775 scope.go:117] "RemoveContainer" containerID="8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e" Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.019707 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vl6j8" Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.047665 4775 scope.go:117] "RemoveContainer" containerID="8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e" Nov 26 06:36:14 crc kubenswrapper[4775]: E1126 06:36:14.048201 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e\": container with ID starting with 8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e not found: ID does not exist" containerID="8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e" Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.048267 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e"} err="failed to get container status \"8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e\": rpc error: code = NotFound desc = could not find container \"8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e\": container with ID starting with 8d342eef8f25e5372487a95b199913034582e23d67b6efb41630bdcd179ca24e not found: ID does not exist" Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.069607 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vl6j8"] Nov 26 06:36:14 crc kubenswrapper[4775]: I1126 06:36:14.077420 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vl6j8"] Nov 26 06:36:15 crc kubenswrapper[4775]: I1126 06:36:15.342628 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" path="/var/lib/kubelet/pods/268d85ad-f165-4cd2-931f-1da2caefe3cd/volumes" Nov 26 06:36:16 crc kubenswrapper[4775]: I1126 06:36:16.038943 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerID="ed87617810ef68e5f7f9bc82fc9681623d3d5cf89c54cb43692664a5b63d5d4f" exitCode=0 Nov 26 06:36:16 crc kubenswrapper[4775]: I1126 06:36:16.039010 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" event={"ID":"b4e29c14-21e8-4dc0-a701-493ae7fbda58","Type":"ContainerDied","Data":"ed87617810ef68e5f7f9bc82fc9681623d3d5cf89c54cb43692664a5b63d5d4f"} Nov 26 06:36:17 crc kubenswrapper[4775]: I1126 06:36:17.049920 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerID="5219bf508a55b97be3b4462a0fc73b952ab0481160f57c32e7538f71be238aeb" exitCode=0 Nov 26 06:36:17 crc kubenswrapper[4775]: I1126 06:36:17.050088 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" event={"ID":"b4e29c14-21e8-4dc0-a701-493ae7fbda58","Type":"ContainerDied","Data":"5219bf508a55b97be3b4462a0fc73b952ab0481160f57c32e7538f71be238aeb"} Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.390197 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.534429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-util\") pod \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.534517 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvb82\" (UniqueName: \"kubernetes.io/projected/b4e29c14-21e8-4dc0-a701-493ae7fbda58-kube-api-access-cvb82\") pod \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.534573 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-bundle\") pod \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\" (UID: \"b4e29c14-21e8-4dc0-a701-493ae7fbda58\") " Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.535771 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-bundle" (OuterVolumeSpecName: "bundle") pod "b4e29c14-21e8-4dc0-a701-493ae7fbda58" (UID: "b4e29c14-21e8-4dc0-a701-493ae7fbda58"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.542066 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4e29c14-21e8-4dc0-a701-493ae7fbda58-kube-api-access-cvb82" (OuterVolumeSpecName: "kube-api-access-cvb82") pod "b4e29c14-21e8-4dc0-a701-493ae7fbda58" (UID: "b4e29c14-21e8-4dc0-a701-493ae7fbda58"). InnerVolumeSpecName "kube-api-access-cvb82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.544262 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-util" (OuterVolumeSpecName: "util") pod "b4e29c14-21e8-4dc0-a701-493ae7fbda58" (UID: "b4e29c14-21e8-4dc0-a701-493ae7fbda58"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.635966 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-util\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.636002 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvb82\" (UniqueName: \"kubernetes.io/projected/b4e29c14-21e8-4dc0-a701-493ae7fbda58-kube-api-access-cvb82\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:18 crc kubenswrapper[4775]: I1126 06:36:18.636016 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b4e29c14-21e8-4dc0-a701-493ae7fbda58-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:36:19 crc kubenswrapper[4775]: I1126 06:36:19.070346 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" event={"ID":"b4e29c14-21e8-4dc0-a701-493ae7fbda58","Type":"ContainerDied","Data":"6eb601b644fa7a600b970f7aa5ab79c4fe4606fc97043e41a32561aeb86eafb0"} Nov 26 06:36:19 crc kubenswrapper[4775]: I1126 06:36:19.070403 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eb601b644fa7a600b970f7aa5ab79c4fe4606fc97043e41a32561aeb86eafb0" Nov 26 06:36:19 crc kubenswrapper[4775]: I1126 06:36:19.070485 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.922373 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r"] Nov 26 06:36:29 crc kubenswrapper[4775]: E1126 06:36:29.923099 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerName="console" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923111 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerName="console" Nov 26 06:36:29 crc kubenswrapper[4775]: E1126 06:36:29.923125 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="util" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923131 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="util" Nov 26 06:36:29 crc kubenswrapper[4775]: E1126 06:36:29.923137 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="extract" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923146 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="extract" Nov 26 06:36:29 crc kubenswrapper[4775]: E1126 06:36:29.923156 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="pull" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923161 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="pull" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923246 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="268d85ad-f165-4cd2-931f-1da2caefe3cd" containerName="console" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923259 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4e29c14-21e8-4dc0-a701-493ae7fbda58" containerName="extract" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.923614 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.925676 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.927013 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.927346 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9snqp" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.927881 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.927961 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 26 06:36:29 crc kubenswrapper[4775]: I1126 06:36:29.936203 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r"] Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.084533 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75xhr\" (UniqueName: \"kubernetes.io/projected/8e93463e-dc10-48b2-b359-494a53052995-kube-api-access-75xhr\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.084864 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e93463e-dc10-48b2-b359-494a53052995-apiservice-cert\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.084990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e93463e-dc10-48b2-b359-494a53052995-webhook-cert\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.161174 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll"] Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.168105 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.172761 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.173093 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.173484 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-sl6ps" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.179709 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll"] Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.185963 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e93463e-dc10-48b2-b359-494a53052995-apiservice-cert\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.186056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e93463e-dc10-48b2-b359-494a53052995-webhook-cert\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.186110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75xhr\" (UniqueName: \"kubernetes.io/projected/8e93463e-dc10-48b2-b359-494a53052995-kube-api-access-75xhr\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.202031 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e93463e-dc10-48b2-b359-494a53052995-webhook-cert\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.209324 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75xhr\" (UniqueName: \"kubernetes.io/projected/8e93463e-dc10-48b2-b359-494a53052995-kube-api-access-75xhr\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.210385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e93463e-dc10-48b2-b359-494a53052995-apiservice-cert\") pod \"metallb-operator-controller-manager-79f98797bf-xz98r\" (UID: \"8e93463e-dc10-48b2-b359-494a53052995\") " pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.241941 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.287269 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0758bd1-82f1-4a2f-a009-75975267444c-webhook-cert\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.287329 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l7kr\" (UniqueName: \"kubernetes.io/projected/e0758bd1-82f1-4a2f-a009-75975267444c-kube-api-access-2l7kr\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.287373 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0758bd1-82f1-4a2f-a009-75975267444c-apiservice-cert\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.388279 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0758bd1-82f1-4a2f-a009-75975267444c-webhook-cert\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.388326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l7kr\" (UniqueName: \"kubernetes.io/projected/e0758bd1-82f1-4a2f-a009-75975267444c-kube-api-access-2l7kr\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.388373 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0758bd1-82f1-4a2f-a009-75975267444c-apiservice-cert\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.401386 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e0758bd1-82f1-4a2f-a009-75975267444c-webhook-cert\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.406369 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e0758bd1-82f1-4a2f-a009-75975267444c-apiservice-cert\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.408192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l7kr\" (UniqueName: \"kubernetes.io/projected/e0758bd1-82f1-4a2f-a009-75975267444c-kube-api-access-2l7kr\") pod \"metallb-operator-webhook-server-766d5f58fc-d4gll\" (UID: \"e0758bd1-82f1-4a2f-a009-75975267444c\") " pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.486306 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.690334 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r"] Nov 26 06:36:30 crc kubenswrapper[4775]: I1126 06:36:30.755903 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll"] Nov 26 06:36:30 crc kubenswrapper[4775]: W1126 06:36:30.757995 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0758bd1_82f1_4a2f_a009_75975267444c.slice/crio-a15264fa1f6cdd6f91009ae89a34b1e92bdfd24e73108605a271541b0f3694fa WatchSource:0}: Error finding container a15264fa1f6cdd6f91009ae89a34b1e92bdfd24e73108605a271541b0f3694fa: Status 404 returned error can't find the container with id a15264fa1f6cdd6f91009ae89a34b1e92bdfd24e73108605a271541b0f3694fa Nov 26 06:36:31 crc kubenswrapper[4775]: I1126 06:36:31.157981 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" event={"ID":"8e93463e-dc10-48b2-b359-494a53052995","Type":"ContainerStarted","Data":"7734f6afbffc68d40aad1164325d22f037a1894a5d235950ecd8d64865d055e8"} Nov 26 06:36:31 crc kubenswrapper[4775]: I1126 06:36:31.160971 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" event={"ID":"e0758bd1-82f1-4a2f-a009-75975267444c","Type":"ContainerStarted","Data":"a15264fa1f6cdd6f91009ae89a34b1e92bdfd24e73108605a271541b0f3694fa"} Nov 26 06:36:36 crc kubenswrapper[4775]: I1126 06:36:36.191092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" event={"ID":"8e93463e-dc10-48b2-b359-494a53052995","Type":"ContainerStarted","Data":"3379dd84c1e29a3053b7d8a9628811c3f9eaad4ba52922166bfdffe418b36b30"} Nov 26 06:36:36 crc kubenswrapper[4775]: I1126 06:36:36.191837 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:36:36 crc kubenswrapper[4775]: I1126 06:36:36.195226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" event={"ID":"e0758bd1-82f1-4a2f-a009-75975267444c","Type":"ContainerStarted","Data":"85b3acc27ee11de390d9e8448525fb3c0b02a26b3c60ace2286390280b9a3aa4"} Nov 26 06:36:36 crc kubenswrapper[4775]: I1126 06:36:36.195522 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:36 crc kubenswrapper[4775]: I1126 06:36:36.229576 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" podStartSLOduration=2.872299998 podStartE2EDuration="7.229544863s" podCreationTimestamp="2025-11-26 06:36:29 +0000 UTC" firstStartedPulling="2025-11-26 06:36:30.706028348 +0000 UTC m=+674.067332300" lastFinishedPulling="2025-11-26 06:36:35.063273193 +0000 UTC m=+678.424577165" observedRunningTime="2025-11-26 06:36:36.22427381 +0000 UTC m=+679.585577772" watchObservedRunningTime="2025-11-26 06:36:36.229544863 +0000 UTC m=+679.590848855" Nov 26 06:36:36 crc kubenswrapper[4775]: I1126 06:36:36.270034 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" podStartSLOduration=1.950089133 podStartE2EDuration="6.270010344s" podCreationTimestamp="2025-11-26 06:36:30 +0000 UTC" firstStartedPulling="2025-11-26 06:36:30.760973022 +0000 UTC m=+674.122276974" lastFinishedPulling="2025-11-26 06:36:35.080894213 +0000 UTC m=+678.442198185" observedRunningTime="2025-11-26 06:36:36.262971002 +0000 UTC m=+679.624274984" watchObservedRunningTime="2025-11-26 06:36:36.270010344 +0000 UTC m=+679.631314306" Nov 26 06:36:50 crc kubenswrapper[4775]: I1126 06:36:50.496433 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-766d5f58fc-d4gll" Nov 26 06:36:51 crc kubenswrapper[4775]: I1126 06:36:51.420389 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:36:51 crc kubenswrapper[4775]: I1126 06:36:51.420525 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:37:10 crc kubenswrapper[4775]: I1126 06:37:10.245031 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-79f98797bf-xz98r" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.014190 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-r5p7b"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.017105 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.019994 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4db9x" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.020628 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.020954 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.021492 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.026209 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.031414 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.033125 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.091588 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-42hxm"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.092411 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.097682 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-jjq76" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.098359 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.098388 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.098359 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.112564 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-lnhds"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.113573 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.115543 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149194 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dccc8029-10ed-427e-a26c-74427323350b-cert\") pod \"frr-k8s-webhook-server-6998585d5-r6bx2\" (UID: \"dccc8029-10ed-427e-a26c-74427323350b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-sockets\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149303 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7gvp\" (UniqueName: \"kubernetes.io/projected/dccc8029-10ed-427e-a26c-74427323350b-kube-api-access-l7gvp\") pod \"frr-k8s-webhook-server-6998585d5-r6bx2\" (UID: \"dccc8029-10ed-427e-a26c-74427323350b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149343 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6d95a8e-b6ca-49a7-9344-872cddc08f75-metrics-certs\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149364 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llr49\" (UniqueName: \"kubernetes.io/projected/d6d95a8e-b6ca-49a7-9344-872cddc08f75-kube-api-access-llr49\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149395 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-metrics\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149421 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-startup\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149446 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-reloader\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.149476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-conf\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.151152 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-lnhds"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6d95a8e-b6ca-49a7-9344-872cddc08f75-metrics-certs\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llr49\" (UniqueName: \"kubernetes.io/projected/d6d95a8e-b6ca-49a7-9344-872cddc08f75-kube-api-access-llr49\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250408 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-metrics-certs\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-metrics\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-startup\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/953ed086-d444-4c86-9217-134ca8126184-metallb-excludel2\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250492 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-reloader\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250517 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-conf\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250538 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dccc8029-10ed-427e-a26c-74427323350b-cert\") pod \"frr-k8s-webhook-server-6998585d5-r6bx2\" (UID: \"dccc8029-10ed-427e-a26c-74427323350b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250566 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkmrb\" (UniqueName: \"kubernetes.io/projected/2f366115-1210-4d05-9f99-b00eede5cee4-kube-api-access-zkmrb\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250604 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-metrics-certs\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250622 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96trm\" (UniqueName: \"kubernetes.io/projected/953ed086-d444-4c86-9217-134ca8126184-kube-api-access-96trm\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250658 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-sockets\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7gvp\" (UniqueName: \"kubernetes.io/projected/dccc8029-10ed-427e-a26c-74427323350b-kube-api-access-l7gvp\") pod \"frr-k8s-webhook-server-6998585d5-r6bx2\" (UID: \"dccc8029-10ed-427e-a26c-74427323350b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.250688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-cert\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.251192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-metrics\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.251400 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-sockets\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.251669 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-startup\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.251996 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-reloader\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.252239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d6d95a8e-b6ca-49a7-9344-872cddc08f75-frr-conf\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.256775 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dccc8029-10ed-427e-a26c-74427323350b-cert\") pod \"frr-k8s-webhook-server-6998585d5-r6bx2\" (UID: \"dccc8029-10ed-427e-a26c-74427323350b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.267332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d6d95a8e-b6ca-49a7-9344-872cddc08f75-metrics-certs\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.271909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7gvp\" (UniqueName: \"kubernetes.io/projected/dccc8029-10ed-427e-a26c-74427323350b-kube-api-access-l7gvp\") pod \"frr-k8s-webhook-server-6998585d5-r6bx2\" (UID: \"dccc8029-10ed-427e-a26c-74427323350b\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.274014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llr49\" (UniqueName: \"kubernetes.io/projected/d6d95a8e-b6ca-49a7-9344-872cddc08f75-kube-api-access-llr49\") pod \"frr-k8s-r5p7b\" (UID: \"d6d95a8e-b6ca-49a7-9344-872cddc08f75\") " pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.351034 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.351428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkmrb\" (UniqueName: \"kubernetes.io/projected/2f366115-1210-4d05-9f99-b00eede5cee4-kube-api-access-zkmrb\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.351807 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-metrics-certs\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.351958 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.351982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96trm\" (UniqueName: \"kubernetes.io/projected/953ed086-d444-4c86-9217-134ca8126184-kube-api-access-96trm\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.352072 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-cert\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: E1126 06:37:11.351920 4775 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Nov 26 06:37:11 crc kubenswrapper[4775]: E1126 06:37:11.352163 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-metrics-certs podName:2f366115-1210-4d05-9f99-b00eede5cee4 nodeName:}" failed. No retries permitted until 2025-11-26 06:37:11.852138167 +0000 UTC m=+715.213442109 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-metrics-certs") pod "controller-6c7b4b5f48-lnhds" (UID: "2f366115-1210-4d05-9f99-b00eede5cee4") : secret "controller-certs-secret" not found Nov 26 06:37:11 crc kubenswrapper[4775]: E1126 06:37:11.352027 4775 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 26 06:37:11 crc kubenswrapper[4775]: E1126 06:37:11.352195 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist podName:953ed086-d444-4c86-9217-134ca8126184 nodeName:}" failed. No retries permitted until 2025-11-26 06:37:11.852188638 +0000 UTC m=+715.213492590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist") pod "speaker-42hxm" (UID: "953ed086-d444-4c86-9217-134ca8126184") : secret "metallb-memberlist" not found Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.352618 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-metrics-certs\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.352658 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/953ed086-d444-4c86-9217-134ca8126184-metallb-excludel2\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.353280 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/953ed086-d444-4c86-9217-134ca8126184-metallb-excludel2\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.355082 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-cert\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.357297 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-metrics-certs\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.367456 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.368623 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96trm\" (UniqueName: \"kubernetes.io/projected/953ed086-d444-4c86-9217-134ca8126184-kube-api-access-96trm\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.376888 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkmrb\" (UniqueName: \"kubernetes.io/projected/2f366115-1210-4d05-9f99-b00eede5cee4-kube-api-access-zkmrb\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.856681 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2"] Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.859210 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-metrics-certs\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.859280 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:11 crc kubenswrapper[4775]: E1126 06:37:11.859518 4775 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 26 06:37:11 crc kubenswrapper[4775]: E1126 06:37:11.859603 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist podName:953ed086-d444-4c86-9217-134ca8126184 nodeName:}" failed. No retries permitted until 2025-11-26 06:37:12.859581288 +0000 UTC m=+716.220885250 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist") pod "speaker-42hxm" (UID: "953ed086-d444-4c86-9217-134ca8126184") : secret "metallb-memberlist" not found Nov 26 06:37:11 crc kubenswrapper[4775]: I1126 06:37:11.865502 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f366115-1210-4d05-9f99-b00eede5cee4-metrics-certs\") pod \"controller-6c7b4b5f48-lnhds\" (UID: \"2f366115-1210-4d05-9f99-b00eede5cee4\") " pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:11 crc kubenswrapper[4775]: W1126 06:37:11.868699 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddccc8029_10ed_427e_a26c_74427323350b.slice/crio-5ee57ff8e37805695d0cd20fe6b04e8f5ad36ab15394ea00674723dee8844df7 WatchSource:0}: Error finding container 5ee57ff8e37805695d0cd20fe6b04e8f5ad36ab15394ea00674723dee8844df7: Status 404 returned error can't find the container with id 5ee57ff8e37805695d0cd20fe6b04e8f5ad36ab15394ea00674723dee8844df7 Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.035913 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.225330 4775 patch_prober.go:28] interesting pod/dns-default-9cq9z container/dns namespace/openshift-dns: Readiness probe status=failure output="Get \"http://10.217.0.42:8181/ready\": dial tcp 10.217.0.42:8181: connect: connection refused" start-of-body= Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.225645 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-dns/dns-default-9cq9z" podUID="a11b6fbe-aabd-45e3-9302-3b3678bba644" containerName="dns" probeResult="failure" output="Get \"http://10.217.0.42:8181/ready\": dial tcp 10.217.0.42:8181: connect: connection refused" Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.472727 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"7c09b509d26414f01ce698ec04cfde729dabe3c9e897010cec428d0cecba9b93"} Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.475449 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" event={"ID":"dccc8029-10ed-427e-a26c-74427323350b","Type":"ContainerStarted","Data":"5ee57ff8e37805695d0cd20fe6b04e8f5ad36ab15394ea00674723dee8844df7"} Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.499120 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-lnhds"] Nov 26 06:37:12 crc kubenswrapper[4775]: W1126 06:37:12.515518 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f366115_1210_4d05_9f99_b00eede5cee4.slice/crio-50a7ff7f259a413ba2ceb07bd54e64a732aa24e8671aa41603aa5a741fc65a8d WatchSource:0}: Error finding container 50a7ff7f259a413ba2ceb07bd54e64a732aa24e8671aa41603aa5a741fc65a8d: Status 404 returned error can't find the container with id 50a7ff7f259a413ba2ceb07bd54e64a732aa24e8671aa41603aa5a741fc65a8d Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.872501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.878535 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/953ed086-d444-4c86-9217-134ca8126184-memberlist\") pod \"speaker-42hxm\" (UID: \"953ed086-d444-4c86-9217-134ca8126184\") " pod="metallb-system/speaker-42hxm" Nov 26 06:37:12 crc kubenswrapper[4775]: I1126 06:37:12.910906 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-42hxm" Nov 26 06:37:12 crc kubenswrapper[4775]: W1126 06:37:12.932214 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod953ed086_d444_4c86_9217_134ca8126184.slice/crio-3aaf80ccb2c4f4764675fb9b8c8a822981a980cd33cd43e01135a025c508a56b WatchSource:0}: Error finding container 3aaf80ccb2c4f4764675fb9b8c8a822981a980cd33cd43e01135a025c508a56b: Status 404 returned error can't find the container with id 3aaf80ccb2c4f4764675fb9b8c8a822981a980cd33cd43e01135a025c508a56b Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.481689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-42hxm" event={"ID":"953ed086-d444-4c86-9217-134ca8126184","Type":"ContainerStarted","Data":"ec10aba2f939cd48e6b2a8754ecb7572b24283fa89ddab1142e7925a0302a173"} Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.482069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-42hxm" event={"ID":"953ed086-d444-4c86-9217-134ca8126184","Type":"ContainerStarted","Data":"80d41ba27d11733420c0ace59223adca6f0482a869a2d83f3d57bfa884e9d6fe"} Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.482081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-42hxm" event={"ID":"953ed086-d444-4c86-9217-134ca8126184","Type":"ContainerStarted","Data":"3aaf80ccb2c4f4764675fb9b8c8a822981a980cd33cd43e01135a025c508a56b"} Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.482251 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-42hxm" Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.483047 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-lnhds" event={"ID":"2f366115-1210-4d05-9f99-b00eede5cee4","Type":"ContainerStarted","Data":"96abda45b521cf217ca00c9b46348aad66d1d5c32b4303e9226b1404b03cb6c4"} Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.483069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-lnhds" event={"ID":"2f366115-1210-4d05-9f99-b00eede5cee4","Type":"ContainerStarted","Data":"b58af3e24a9f6dbfee6513ad1d1daea373d14e8678e7637bb283437865a52fa3"} Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.483079 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-lnhds" event={"ID":"2f366115-1210-4d05-9f99-b00eede5cee4","Type":"ContainerStarted","Data":"50a7ff7f259a413ba2ceb07bd54e64a732aa24e8671aa41603aa5a741fc65a8d"} Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.483376 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.500241 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-42hxm" podStartSLOduration=2.500220788 podStartE2EDuration="2.500220788s" podCreationTimestamp="2025-11-26 06:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:37:13.498375418 +0000 UTC m=+716.859679380" watchObservedRunningTime="2025-11-26 06:37:13.500220788 +0000 UTC m=+716.861524740" Nov 26 06:37:13 crc kubenswrapper[4775]: I1126 06:37:13.516991 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-lnhds" podStartSLOduration=2.516971384 podStartE2EDuration="2.516971384s" podCreationTimestamp="2025-11-26 06:37:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:37:13.514849526 +0000 UTC m=+716.876153478" watchObservedRunningTime="2025-11-26 06:37:13.516971384 +0000 UTC m=+716.878275336" Nov 26 06:37:17 crc kubenswrapper[4775]: I1126 06:37:17.761929 4775 scope.go:117] "RemoveContainer" containerID="4051a1dda93f85ff24534d040d245e89877b6b43675359ff835667ed9945f446" Nov 26 06:37:19 crc kubenswrapper[4775]: I1126 06:37:19.523106 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6d95a8e-b6ca-49a7-9344-872cddc08f75" containerID="c5e1b5594c11c4579fe3a14ec1dc9c6f9e73b5762ca6edfeffcf33349309fa4d" exitCode=0 Nov 26 06:37:19 crc kubenswrapper[4775]: I1126 06:37:19.523208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerDied","Data":"c5e1b5594c11c4579fe3a14ec1dc9c6f9e73b5762ca6edfeffcf33349309fa4d"} Nov 26 06:37:19 crc kubenswrapper[4775]: I1126 06:37:19.525752 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" event={"ID":"dccc8029-10ed-427e-a26c-74427323350b","Type":"ContainerStarted","Data":"33ad782db0932381bb36c09f66f6872e113a59f73dc8f8e30e883ff7eb1850db"} Nov 26 06:37:19 crc kubenswrapper[4775]: I1126 06:37:19.526417 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:19 crc kubenswrapper[4775]: I1126 06:37:19.589061 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" podStartSLOduration=3.048071133 podStartE2EDuration="9.589030259s" podCreationTimestamp="2025-11-26 06:37:10 +0000 UTC" firstStartedPulling="2025-11-26 06:37:11.872917861 +0000 UTC m=+715.234221813" lastFinishedPulling="2025-11-26 06:37:18.413876967 +0000 UTC m=+721.775180939" observedRunningTime="2025-11-26 06:37:19.581426182 +0000 UTC m=+722.942730214" watchObservedRunningTime="2025-11-26 06:37:19.589030259 +0000 UTC m=+722.950334251" Nov 26 06:37:20 crc kubenswrapper[4775]: I1126 06:37:20.534001 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6d95a8e-b6ca-49a7-9344-872cddc08f75" containerID="8f40f3fc016368d4bb0116f8a1a9de2f18522f471b3d4b73824db36bc7c38cd7" exitCode=0 Nov 26 06:37:20 crc kubenswrapper[4775]: I1126 06:37:20.534085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerDied","Data":"8f40f3fc016368d4bb0116f8a1a9de2f18522f471b3d4b73824db36bc7c38cd7"} Nov 26 06:37:21 crc kubenswrapper[4775]: I1126 06:37:21.420365 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:37:21 crc kubenswrapper[4775]: I1126 06:37:21.420450 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:37:21 crc kubenswrapper[4775]: I1126 06:37:21.543572 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6d95a8e-b6ca-49a7-9344-872cddc08f75" containerID="a5003b32887df9147547e7c517ecac0c4bf0bf583893a552021a0031a2e07e55" exitCode=0 Nov 26 06:37:21 crc kubenswrapper[4775]: I1126 06:37:21.543675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerDied","Data":"a5003b32887df9147547e7c517ecac0c4bf0bf583893a552021a0031a2e07e55"} Nov 26 06:37:22 crc kubenswrapper[4775]: I1126 06:37:22.043044 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-lnhds" Nov 26 06:37:22 crc kubenswrapper[4775]: I1126 06:37:22.557969 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"ca18a3222fd5c168297ff8c44e1d5260237819563f45c2a6af0bc234918cf6e6"} Nov 26 06:37:22 crc kubenswrapper[4775]: I1126 06:37:22.558027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"1ad9a7f52d877c3b2efce2a812b0a40fc0f7dc82a725779074857f4cd709394b"} Nov 26 06:37:22 crc kubenswrapper[4775]: I1126 06:37:22.558036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"7ba49a9f1dd876800e8f6e5f42cf81e78860809ce955c9b1c419843942cf8ca2"} Nov 26 06:37:22 crc kubenswrapper[4775]: I1126 06:37:22.558044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"f9adc8d6e23ee0b3a16f4a1ba13ecd5ce28e19ca47ca559d78e1f6249126d939"} Nov 26 06:37:22 crc kubenswrapper[4775]: I1126 06:37:22.558052 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"8680e343457839259970252a475bcd662ed0222be7eed16571641890c05c4313"} Nov 26 06:37:23 crc kubenswrapper[4775]: I1126 06:37:23.573397 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-r5p7b" event={"ID":"d6d95a8e-b6ca-49a7-9344-872cddc08f75","Type":"ContainerStarted","Data":"887bf7170911edfb162d223356ae93669f204af96486b88746f72d65b8c77615"} Nov 26 06:37:23 crc kubenswrapper[4775]: I1126 06:37:23.573946 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:23 crc kubenswrapper[4775]: I1126 06:37:23.613693 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-r5p7b" podStartSLOduration=6.74800777 podStartE2EDuration="13.613665768s" podCreationTimestamp="2025-11-26 06:37:10 +0000 UTC" firstStartedPulling="2025-11-26 06:37:11.528266937 +0000 UTC m=+714.889570929" lastFinishedPulling="2025-11-26 06:37:18.393924975 +0000 UTC m=+721.755228927" observedRunningTime="2025-11-26 06:37:23.609060063 +0000 UTC m=+726.970364085" watchObservedRunningTime="2025-11-26 06:37:23.613665768 +0000 UTC m=+726.974969800" Nov 26 06:37:26 crc kubenswrapper[4775]: I1126 06:37:26.351544 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:26 crc kubenswrapper[4775]: I1126 06:37:26.413388 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:31 crc kubenswrapper[4775]: I1126 06:37:31.359010 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-r5p7b" Nov 26 06:37:31 crc kubenswrapper[4775]: I1126 06:37:31.377508 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-r6bx2" Nov 26 06:37:32 crc kubenswrapper[4775]: I1126 06:37:32.915657 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-42hxm" Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.816824 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fwgf9"] Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.818838 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.822459 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-jssrg" Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.822835 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.855648 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.900059 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fwgf9"] Nov 26 06:37:35 crc kubenswrapper[4775]: I1126 06:37:35.927510 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hx6n\" (UniqueName: \"kubernetes.io/projected/81158087-b698-4505-8955-6ec2f823a90d-kube-api-access-7hx6n\") pod \"openstack-operator-index-fwgf9\" (UID: \"81158087-b698-4505-8955-6ec2f823a90d\") " pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:36 crc kubenswrapper[4775]: I1126 06:37:36.029548 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hx6n\" (UniqueName: \"kubernetes.io/projected/81158087-b698-4505-8955-6ec2f823a90d-kube-api-access-7hx6n\") pod \"openstack-operator-index-fwgf9\" (UID: \"81158087-b698-4505-8955-6ec2f823a90d\") " pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:36 crc kubenswrapper[4775]: I1126 06:37:36.045540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hx6n\" (UniqueName: \"kubernetes.io/projected/81158087-b698-4505-8955-6ec2f823a90d-kube-api-access-7hx6n\") pod \"openstack-operator-index-fwgf9\" (UID: \"81158087-b698-4505-8955-6ec2f823a90d\") " pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:36 crc kubenswrapper[4775]: I1126 06:37:36.176827 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:36 crc kubenswrapper[4775]: I1126 06:37:36.646949 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fwgf9"] Nov 26 06:37:36 crc kubenswrapper[4775]: W1126 06:37:36.653764 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81158087_b698_4505_8955_6ec2f823a90d.slice/crio-6c5491b46799b613b8d47935054aca5b548d246df5d9adb5447df704bd66dbaa WatchSource:0}: Error finding container 6c5491b46799b613b8d47935054aca5b548d246df5d9adb5447df704bd66dbaa: Status 404 returned error can't find the container with id 6c5491b46799b613b8d47935054aca5b548d246df5d9adb5447df704bd66dbaa Nov 26 06:37:36 crc kubenswrapper[4775]: I1126 06:37:36.716268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwgf9" event={"ID":"81158087-b698-4505-8955-6ec2f823a90d","Type":"ContainerStarted","Data":"6c5491b46799b613b8d47935054aca5b548d246df5d9adb5447df704bd66dbaa"} Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.172371 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fwgf9"] Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.735845 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwgf9" event={"ID":"81158087-b698-4505-8955-6ec2f823a90d","Type":"ContainerStarted","Data":"f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0"} Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.758494 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fwgf9" podStartSLOduration=2.428747533 podStartE2EDuration="4.758470135s" podCreationTimestamp="2025-11-26 06:37:35 +0000 UTC" firstStartedPulling="2025-11-26 06:37:36.657528887 +0000 UTC m=+740.018832879" lastFinishedPulling="2025-11-26 06:37:38.987251489 +0000 UTC m=+742.348555481" observedRunningTime="2025-11-26 06:37:39.753414667 +0000 UTC m=+743.114718659" watchObservedRunningTime="2025-11-26 06:37:39.758470135 +0000 UTC m=+743.119774147" Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.788656 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rpzlv"] Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.789897 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.800116 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rpzlv"] Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.884371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4c46\" (UniqueName: \"kubernetes.io/projected/cd5862a0-b9ab-4473-bbe4-34253c8cbb79-kube-api-access-g4c46\") pod \"openstack-operator-index-rpzlv\" (UID: \"cd5862a0-b9ab-4473-bbe4-34253c8cbb79\") " pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:39 crc kubenswrapper[4775]: I1126 06:37:39.985542 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4c46\" (UniqueName: \"kubernetes.io/projected/cd5862a0-b9ab-4473-bbe4-34253c8cbb79-kube-api-access-g4c46\") pod \"openstack-operator-index-rpzlv\" (UID: \"cd5862a0-b9ab-4473-bbe4-34253c8cbb79\") " pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:40 crc kubenswrapper[4775]: I1126 06:37:40.023320 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4c46\" (UniqueName: \"kubernetes.io/projected/cd5862a0-b9ab-4473-bbe4-34253c8cbb79-kube-api-access-g4c46\") pod \"openstack-operator-index-rpzlv\" (UID: \"cd5862a0-b9ab-4473-bbe4-34253c8cbb79\") " pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:40 crc kubenswrapper[4775]: I1126 06:37:40.124037 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:40 crc kubenswrapper[4775]: I1126 06:37:40.621542 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rpzlv"] Nov 26 06:37:40 crc kubenswrapper[4775]: W1126 06:37:40.627687 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd5862a0_b9ab_4473_bbe4_34253c8cbb79.slice/crio-1a19bfb24de9f989774fdb21a66f8b2b2a51a1dbf6cd480949429ea8d894588b WatchSource:0}: Error finding container 1a19bfb24de9f989774fdb21a66f8b2b2a51a1dbf6cd480949429ea8d894588b: Status 404 returned error can't find the container with id 1a19bfb24de9f989774fdb21a66f8b2b2a51a1dbf6cd480949429ea8d894588b Nov 26 06:37:40 crc kubenswrapper[4775]: I1126 06:37:40.743375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rpzlv" event={"ID":"cd5862a0-b9ab-4473-bbe4-34253c8cbb79","Type":"ContainerStarted","Data":"1a19bfb24de9f989774fdb21a66f8b2b2a51a1dbf6cd480949429ea8d894588b"} Nov 26 06:37:40 crc kubenswrapper[4775]: I1126 06:37:40.743497 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-fwgf9" podUID="81158087-b698-4505-8955-6ec2f823a90d" containerName="registry-server" containerID="cri-o://f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0" gracePeriod=2 Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.266255 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.406236 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hx6n\" (UniqueName: \"kubernetes.io/projected/81158087-b698-4505-8955-6ec2f823a90d-kube-api-access-7hx6n\") pod \"81158087-b698-4505-8955-6ec2f823a90d\" (UID: \"81158087-b698-4505-8955-6ec2f823a90d\") " Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.411928 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81158087-b698-4505-8955-6ec2f823a90d-kube-api-access-7hx6n" (OuterVolumeSpecName: "kube-api-access-7hx6n") pod "81158087-b698-4505-8955-6ec2f823a90d" (UID: "81158087-b698-4505-8955-6ec2f823a90d"). InnerVolumeSpecName "kube-api-access-7hx6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.508809 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hx6n\" (UniqueName: \"kubernetes.io/projected/81158087-b698-4505-8955-6ec2f823a90d-kube-api-access-7hx6n\") on node \"crc\" DevicePath \"\"" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.755487 4775 generic.go:334] "Generic (PLEG): container finished" podID="81158087-b698-4505-8955-6ec2f823a90d" containerID="f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0" exitCode=0 Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.755544 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwgf9" event={"ID":"81158087-b698-4505-8955-6ec2f823a90d","Type":"ContainerDied","Data":"f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0"} Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.755654 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fwgf9" event={"ID":"81158087-b698-4505-8955-6ec2f823a90d","Type":"ContainerDied","Data":"6c5491b46799b613b8d47935054aca5b548d246df5d9adb5447df704bd66dbaa"} Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.755686 4775 scope.go:117] "RemoveContainer" containerID="f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.758071 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rpzlv" event={"ID":"cd5862a0-b9ab-4473-bbe4-34253c8cbb79","Type":"ContainerStarted","Data":"c64a31e8ef567c6a6a8d989b2aa88040f5b42a12f2c4902d4cc4aca266dfc336"} Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.758778 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fwgf9" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.790429 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rpzlv" podStartSLOduration=2.736508082 podStartE2EDuration="2.790410038s" podCreationTimestamp="2025-11-26 06:37:39 +0000 UTC" firstStartedPulling="2025-11-26 06:37:40.634451779 +0000 UTC m=+743.995755731" lastFinishedPulling="2025-11-26 06:37:40.688353705 +0000 UTC m=+744.049657687" observedRunningTime="2025-11-26 06:37:41.77874499 +0000 UTC m=+745.140048982" watchObservedRunningTime="2025-11-26 06:37:41.790410038 +0000 UTC m=+745.151714000" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.793319 4775 scope.go:117] "RemoveContainer" containerID="f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0" Nov 26 06:37:41 crc kubenswrapper[4775]: E1126 06:37:41.794072 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0\": container with ID starting with f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0 not found: ID does not exist" containerID="f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.794113 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0"} err="failed to get container status \"f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0\": rpc error: code = NotFound desc = could not find container \"f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0\": container with ID starting with f771645152c2c9714e7b18a462ba57e1eafb9a4dbab19d0fc547c41e0a3cccc0 not found: ID does not exist" Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.818543 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fwgf9"] Nov 26 06:37:41 crc kubenswrapper[4775]: I1126 06:37:41.826873 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-fwgf9"] Nov 26 06:37:43 crc kubenswrapper[4775]: I1126 06:37:43.337907 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81158087-b698-4505-8955-6ec2f823a90d" path="/var/lib/kubelet/pods/81158087-b698-4505-8955-6ec2f823a90d/volumes" Nov 26 06:37:48 crc kubenswrapper[4775]: I1126 06:37:48.382303 4775 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 06:37:50 crc kubenswrapper[4775]: I1126 06:37:50.124292 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:50 crc kubenswrapper[4775]: I1126 06:37:50.124745 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:50 crc kubenswrapper[4775]: I1126 06:37:50.166348 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:50 crc kubenswrapper[4775]: I1126 06:37:50.891774 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-rpzlv" Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.419588 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.419985 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.420075 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.421084 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8685d35e684f235bf0161f8f3e1dacf8484e59a0d24a51ad51c239999335747e"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.421177 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://8685d35e684f235bf0161f8f3e1dacf8484e59a0d24a51ad51c239999335747e" gracePeriod=600 Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.866032 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="8685d35e684f235bf0161f8f3e1dacf8484e59a0d24a51ad51c239999335747e" exitCode=0 Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.866127 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"8685d35e684f235bf0161f8f3e1dacf8484e59a0d24a51ad51c239999335747e"} Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.866522 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"c257f11ab63c2232fe4ea03480103fc149650fe9c5d870044eeec21f9ccaaf75"} Nov 26 06:37:51 crc kubenswrapper[4775]: I1126 06:37:51.866554 4775 scope.go:117] "RemoveContainer" containerID="988eb2b1e284ef66703f1f37b2da654fd669430dce2be08f52d009293283f65e" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.636453 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff"] Nov 26 06:37:55 crc kubenswrapper[4775]: E1126 06:37:55.637820 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81158087-b698-4505-8955-6ec2f823a90d" containerName="registry-server" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.637854 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="81158087-b698-4505-8955-6ec2f823a90d" containerName="registry-server" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.638152 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="81158087-b698-4505-8955-6ec2f823a90d" containerName="registry-server" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.640038 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.646286 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-ch7sp" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.647658 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff"] Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.837031 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-bundle\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.837130 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sxhc\" (UniqueName: \"kubernetes.io/projected/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-kube-api-access-7sxhc\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.837259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-util\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.939324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sxhc\" (UniqueName: \"kubernetes.io/projected/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-kube-api-access-7sxhc\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.939471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-util\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.939627 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-bundle\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.940380 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-util\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.940534 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-bundle\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:55 crc kubenswrapper[4775]: I1126 06:37:55.974360 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sxhc\" (UniqueName: \"kubernetes.io/projected/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-kube-api-access-7sxhc\") pod \"83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:56 crc kubenswrapper[4775]: I1126 06:37:56.260405 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:37:56 crc kubenswrapper[4775]: I1126 06:37:56.782585 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff"] Nov 26 06:37:56 crc kubenswrapper[4775]: I1126 06:37:56.913892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" event={"ID":"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb","Type":"ContainerStarted","Data":"22c530a14970b6ea166570e689e1909c20720988fde4a13eea0d503b93e900b4"} Nov 26 06:37:57 crc kubenswrapper[4775]: I1126 06:37:57.925523 4775 generic.go:334] "Generic (PLEG): container finished" podID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerID="aea9afba891c6931d1db9f6692a3c9a11151b16e33193496b6a00204b5e68afd" exitCode=0 Nov 26 06:37:57 crc kubenswrapper[4775]: I1126 06:37:57.925596 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" event={"ID":"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb","Type":"ContainerDied","Data":"aea9afba891c6931d1db9f6692a3c9a11151b16e33193496b6a00204b5e68afd"} Nov 26 06:37:58 crc kubenswrapper[4775]: I1126 06:37:58.935617 4775 generic.go:334] "Generic (PLEG): container finished" podID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerID="c00b0248f637391722ea61f669d0946cbf90b4eaaf564248b52f9f67946067e2" exitCode=0 Nov 26 06:37:58 crc kubenswrapper[4775]: I1126 06:37:58.935683 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" event={"ID":"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb","Type":"ContainerDied","Data":"c00b0248f637391722ea61f669d0946cbf90b4eaaf564248b52f9f67946067e2"} Nov 26 06:37:59 crc kubenswrapper[4775]: I1126 06:37:59.946473 4775 generic.go:334] "Generic (PLEG): container finished" podID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerID="c4a759516e28ae96b6e3b55731d85ea360750c6a37245b7c73cd7b5c766e2d89" exitCode=0 Nov 26 06:37:59 crc kubenswrapper[4775]: I1126 06:37:59.946560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" event={"ID":"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb","Type":"ContainerDied","Data":"c4a759516e28ae96b6e3b55731d85ea360750c6a37245b7c73cd7b5c766e2d89"} Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.245543 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.426787 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-util\") pod \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.427175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-bundle\") pod \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.427374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sxhc\" (UniqueName: \"kubernetes.io/projected/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-kube-api-access-7sxhc\") pod \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\" (UID: \"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb\") " Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.427903 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-bundle" (OuterVolumeSpecName: "bundle") pod "78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" (UID: "78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.432384 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-kube-api-access-7sxhc" (OuterVolumeSpecName: "kube-api-access-7sxhc") pod "78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" (UID: "78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb"). InnerVolumeSpecName "kube-api-access-7sxhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.450152 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-util" (OuterVolumeSpecName: "util") pod "78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" (UID: "78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.529539 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sxhc\" (UniqueName: \"kubernetes.io/projected/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-kube-api-access-7sxhc\") on node \"crc\" DevicePath \"\"" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.529573 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-util\") on node \"crc\" DevicePath \"\"" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.529587 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.968560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" event={"ID":"78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb","Type":"ContainerDied","Data":"22c530a14970b6ea166570e689e1909c20720988fde4a13eea0d503b93e900b4"} Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.968615 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c530a14970b6ea166570e689e1909c20720988fde4a13eea0d503b93e900b4" Nov 26 06:38:01 crc kubenswrapper[4775]: I1126 06:38:01.968744 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.861729 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd"] Nov 26 06:38:07 crc kubenswrapper[4775]: E1126 06:38:07.862287 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="util" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.862305 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="util" Nov 26 06:38:07 crc kubenswrapper[4775]: E1126 06:38:07.862322 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="extract" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.862330 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="extract" Nov 26 06:38:07 crc kubenswrapper[4775]: E1126 06:38:07.862348 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="pull" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.862356 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="pull" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.862488 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb" containerName="extract" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.862923 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.865430 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-v6kz7" Nov 26 06:38:07 crc kubenswrapper[4775]: I1126 06:38:07.882568 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd"] Nov 26 06:38:08 crc kubenswrapper[4775]: I1126 06:38:08.021632 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc2kw\" (UniqueName: \"kubernetes.io/projected/2b412dfe-90b0-4018-81ca-32d9644fbc99-kube-api-access-tc2kw\") pod \"openstack-operator-controller-operator-6df95dcb7f-5wnrd\" (UID: \"2b412dfe-90b0-4018-81ca-32d9644fbc99\") " pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:08 crc kubenswrapper[4775]: I1126 06:38:08.122975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc2kw\" (UniqueName: \"kubernetes.io/projected/2b412dfe-90b0-4018-81ca-32d9644fbc99-kube-api-access-tc2kw\") pod \"openstack-operator-controller-operator-6df95dcb7f-5wnrd\" (UID: \"2b412dfe-90b0-4018-81ca-32d9644fbc99\") " pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:08 crc kubenswrapper[4775]: I1126 06:38:08.142406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc2kw\" (UniqueName: \"kubernetes.io/projected/2b412dfe-90b0-4018-81ca-32d9644fbc99-kube-api-access-tc2kw\") pod \"openstack-operator-controller-operator-6df95dcb7f-5wnrd\" (UID: \"2b412dfe-90b0-4018-81ca-32d9644fbc99\") " pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:08 crc kubenswrapper[4775]: I1126 06:38:08.183287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:08 crc kubenswrapper[4775]: I1126 06:38:08.462998 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd"] Nov 26 06:38:08 crc kubenswrapper[4775]: W1126 06:38:08.484808 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b412dfe_90b0_4018_81ca_32d9644fbc99.slice/crio-4a7beda6ea368f18f32beb36317e3432cabaf18dbdc863030835375362efd29b WatchSource:0}: Error finding container 4a7beda6ea368f18f32beb36317e3432cabaf18dbdc863030835375362efd29b: Status 404 returned error can't find the container with id 4a7beda6ea368f18f32beb36317e3432cabaf18dbdc863030835375362efd29b Nov 26 06:38:09 crc kubenswrapper[4775]: I1126 06:38:09.020221 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" event={"ID":"2b412dfe-90b0-4018-81ca-32d9644fbc99","Type":"ContainerStarted","Data":"4a7beda6ea368f18f32beb36317e3432cabaf18dbdc863030835375362efd29b"} Nov 26 06:38:13 crc kubenswrapper[4775]: I1126 06:38:13.056314 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" event={"ID":"2b412dfe-90b0-4018-81ca-32d9644fbc99","Type":"ContainerStarted","Data":"7bcec9b9110fa9059c9cf2e4b654f3e3af96bc6dd6775968e920a604a813aaca"} Nov 26 06:38:13 crc kubenswrapper[4775]: I1126 06:38:13.056968 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:13 crc kubenswrapper[4775]: I1126 06:38:13.094146 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" podStartSLOduration=2.043912443 podStartE2EDuration="6.094129159s" podCreationTimestamp="2025-11-26 06:38:07 +0000 UTC" firstStartedPulling="2025-11-26 06:38:08.486857883 +0000 UTC m=+771.848161845" lastFinishedPulling="2025-11-26 06:38:12.537074599 +0000 UTC m=+775.898378561" observedRunningTime="2025-11-26 06:38:13.093106892 +0000 UTC m=+776.454410894" watchObservedRunningTime="2025-11-26 06:38:13.094129159 +0000 UTC m=+776.455433121" Nov 26 06:38:18 crc kubenswrapper[4775]: I1126 06:38:18.186289 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6df95dcb7f-5wnrd" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.747666 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.749035 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.752185 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-69tk8" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.755230 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.756646 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.759174 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-kt8d5" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.768892 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.774554 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.784108 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-dghfx"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.785248 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.793561 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-kbpb8" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.798428 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-dghfx"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.822007 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.823207 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.828257 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-99w4t" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.838564 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.839782 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.848558 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-9hnfm" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.852597 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.859439 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.889283 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.890528 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.892404 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5smcg" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.898892 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.900061 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.910178 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-97sv5" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.910372 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.920483 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.933822 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.935056 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-555fc\" (UniqueName: \"kubernetes.io/projected/1a1d393e-2d6c-425c-9e6f-068d4c8ee090-kube-api-access-555fc\") pod \"designate-operator-controller-manager-955677c94-dghfx\" (UID: \"1a1d393e-2d6c-425c-9e6f-068d4c8ee090\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.935104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wq82\" (UniqueName: \"kubernetes.io/projected/cc351f29-006f-4ee4-80cb-5860f4f824ff-kube-api-access-8wq82\") pod \"glance-operator-controller-manager-589cbd6b5b-sdz7q\" (UID: \"cc351f29-006f-4ee4-80cb-5860f4f824ff\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.935116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.935151 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhq8c\" (UniqueName: \"kubernetes.io/projected/153b27e4-d948-4ff6-9d6a-bfbcebd17cad-kube-api-access-zhq8c\") pod \"cinder-operator-controller-manager-6b7f75547b-5whjf\" (UID: \"153b27e4-d948-4ff6-9d6a-bfbcebd17cad\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.935205 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkxpl\" (UniqueName: \"kubernetes.io/projected/4f5effa7-3458-4978-8382-0f5b4a17105a-kube-api-access-wkxpl\") pod \"barbican-operator-controller-manager-7b64f4fb85-27cqk\" (UID: \"4f5effa7-3458-4978-8382-0f5b4a17105a\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.943184 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-v8ft6" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.951758 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.985780 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v"] Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.987938 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:38:36 crc kubenswrapper[4775]: I1126 06:38:36.990173 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-rp8m9" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j5kj\" (UniqueName: \"kubernetes.io/projected/8483a9f8-d70d-497d-ae02-c6b426a49306-kube-api-access-8j5kj\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk2kr\" (UniqueName: \"kubernetes.io/projected/faa5a2e8-ca2b-42e1-ac9a-3580111707ac-kube-api-access-sk2kr\") pod \"heat-operator-controller-manager-5b77f656f-d7m5d\" (UID: \"faa5a2e8-ca2b-42e1-ac9a-3580111707ac\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038706 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t98bv\" (UniqueName: \"kubernetes.io/projected/2158b3e8-3c08-4e7c-9b28-cd32534135a1-kube-api-access-t98bv\") pod \"ironic-operator-controller-manager-67cb4dc6d4-gdk4v\" (UID: \"2158b3e8-3c08-4e7c-9b28-cd32534135a1\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038761 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkxpl\" (UniqueName: \"kubernetes.io/projected/4f5effa7-3458-4978-8382-0f5b4a17105a-kube-api-access-wkxpl\") pod \"barbican-operator-controller-manager-7b64f4fb85-27cqk\" (UID: \"4f5effa7-3458-4978-8382-0f5b4a17105a\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038888 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-555fc\" (UniqueName: \"kubernetes.io/projected/1a1d393e-2d6c-425c-9e6f-068d4c8ee090-kube-api-access-555fc\") pod \"designate-operator-controller-manager-955677c94-dghfx\" (UID: \"1a1d393e-2d6c-425c-9e6f-068d4c8ee090\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ntk6\" (UniqueName: \"kubernetes.io/projected/997d00b8-4cc3-45e9-8af2-2d573eef844a-kube-api-access-6ntk6\") pod \"keystone-operator-controller-manager-7b4567c7cf-4fxcp\" (UID: \"997d00b8-4cc3-45e9-8af2-2d573eef844a\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.038995 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wq82\" (UniqueName: \"kubernetes.io/projected/cc351f29-006f-4ee4-80cb-5860f4f824ff-kube-api-access-8wq82\") pod \"glance-operator-controller-manager-589cbd6b5b-sdz7q\" (UID: \"cc351f29-006f-4ee4-80cb-5860f4f824ff\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.039042 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gxv7\" (UniqueName: \"kubernetes.io/projected/ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7-kube-api-access-5gxv7\") pod \"horizon-operator-controller-manager-5d494799bf-jmwr9\" (UID: \"ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.039110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.039140 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.039169 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhq8c\" (UniqueName: \"kubernetes.io/projected/153b27e4-d948-4ff6-9d6a-bfbcebd17cad-kube-api-access-zhq8c\") pod \"cinder-operator-controller-manager-6b7f75547b-5whjf\" (UID: \"153b27e4-d948-4ff6-9d6a-bfbcebd17cad\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.083292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-555fc\" (UniqueName: \"kubernetes.io/projected/1a1d393e-2d6c-425c-9e6f-068d4c8ee090-kube-api-access-555fc\") pod \"designate-operator-controller-manager-955677c94-dghfx\" (UID: \"1a1d393e-2d6c-425c-9e6f-068d4c8ee090\") " pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.085896 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhq8c\" (UniqueName: \"kubernetes.io/projected/153b27e4-d948-4ff6-9d6a-bfbcebd17cad-kube-api-access-zhq8c\") pod \"cinder-operator-controller-manager-6b7f75547b-5whjf\" (UID: \"153b27e4-d948-4ff6-9d6a-bfbcebd17cad\") " pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.086327 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.090365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wq82\" (UniqueName: \"kubernetes.io/projected/cc351f29-006f-4ee4-80cb-5860f4f824ff-kube-api-access-8wq82\") pod \"glance-operator-controller-manager-589cbd6b5b-sdz7q\" (UID: \"cc351f29-006f-4ee4-80cb-5860f4f824ff\") " pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.100908 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.101381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkxpl\" (UniqueName: \"kubernetes.io/projected/4f5effa7-3458-4978-8382-0f5b4a17105a-kube-api-access-wkxpl\") pod \"barbican-operator-controller-manager-7b64f4fb85-27cqk\" (UID: \"4f5effa7-3458-4978-8382-0f5b4a17105a\") " pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.114612 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.115909 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.123171 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rntgq" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.123860 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.134117 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140162 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140217 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf9h9\" (UniqueName: \"kubernetes.io/projected/021e1511-a060-4d30-af62-872ba84b84c3-kube-api-access-hf9h9\") pod \"manila-operator-controller-manager-5d499bf58b-5hvj7\" (UID: \"021e1511-a060-4d30-af62-872ba84b84c3\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140256 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j5kj\" (UniqueName: \"kubernetes.io/projected/8483a9f8-d70d-497d-ae02-c6b426a49306-kube-api-access-8j5kj\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140275 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk2kr\" (UniqueName: \"kubernetes.io/projected/faa5a2e8-ca2b-42e1-ac9a-3580111707ac-kube-api-access-sk2kr\") pod \"heat-operator-controller-manager-5b77f656f-d7m5d\" (UID: \"faa5a2e8-ca2b-42e1-ac9a-3580111707ac\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140299 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t98bv\" (UniqueName: \"kubernetes.io/projected/2158b3e8-3c08-4e7c-9b28-cd32534135a1-kube-api-access-t98bv\") pod \"ironic-operator-controller-manager-67cb4dc6d4-gdk4v\" (UID: \"2158b3e8-3c08-4e7c-9b28-cd32534135a1\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ntk6\" (UniqueName: \"kubernetes.io/projected/997d00b8-4cc3-45e9-8af2-2d573eef844a-kube-api-access-6ntk6\") pod \"keystone-operator-controller-manager-7b4567c7cf-4fxcp\" (UID: \"997d00b8-4cc3-45e9-8af2-2d573eef844a\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.140358 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gxv7\" (UniqueName: \"kubernetes.io/projected/ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7-kube-api-access-5gxv7\") pod \"horizon-operator-controller-manager-5d494799bf-jmwr9\" (UID: \"ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.140883 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.140958 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert podName:8483a9f8-d70d-497d-ae02-c6b426a49306 nodeName:}" failed. No retries permitted until 2025-11-26 06:38:37.64093648 +0000 UTC m=+801.002240512 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert") pod "infra-operator-controller-manager-57548d458d-xrzlh" (UID: "8483a9f8-d70d-497d-ae02-c6b426a49306") : secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.141672 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.145306 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-f8hvj" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.150335 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.156164 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.158334 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.162795 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t98bv\" (UniqueName: \"kubernetes.io/projected/2158b3e8-3c08-4e7c-9b28-cd32534135a1-kube-api-access-t98bv\") pod \"ironic-operator-controller-manager-67cb4dc6d4-gdk4v\" (UID: \"2158b3e8-3c08-4e7c-9b28-cd32534135a1\") " pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.164274 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ntk6\" (UniqueName: \"kubernetes.io/projected/997d00b8-4cc3-45e9-8af2-2d573eef844a-kube-api-access-6ntk6\") pod \"keystone-operator-controller-manager-7b4567c7cf-4fxcp\" (UID: \"997d00b8-4cc3-45e9-8af2-2d573eef844a\") " pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.164326 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.164446 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jnzg9" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.167317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j5kj\" (UniqueName: \"kubernetes.io/projected/8483a9f8-d70d-497d-ae02-c6b426a49306-kube-api-access-8j5kj\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.173900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk2kr\" (UniqueName: \"kubernetes.io/projected/faa5a2e8-ca2b-42e1-ac9a-3580111707ac-kube-api-access-sk2kr\") pod \"heat-operator-controller-manager-5b77f656f-d7m5d\" (UID: \"faa5a2e8-ca2b-42e1-ac9a-3580111707ac\") " pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.174467 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.175683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gxv7\" (UniqueName: \"kubernetes.io/projected/ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7-kube-api-access-5gxv7\") pod \"horizon-operator-controller-manager-5d494799bf-jmwr9\" (UID: \"ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7\") " pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.176293 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.177255 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.181108 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-s4vzd" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.196267 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.203373 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.204407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.207542 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.208500 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-fnb59" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.209660 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.211804 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-2jvjl" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.212214 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.212677 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.215021 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.221473 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.225891 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.232583 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.234428 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.234675 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.237203 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-d2scs" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.237399 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-69gh2" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.241526 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf9h9\" (UniqueName: \"kubernetes.io/projected/021e1511-a060-4d30-af62-872ba84b84c3-kube-api-access-hf9h9\") pod \"manila-operator-controller-manager-5d499bf58b-5hvj7\" (UID: \"021e1511-a060-4d30-af62-872ba84b84c3\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.253071 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.255645 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.261058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf9h9\" (UniqueName: \"kubernetes.io/projected/021e1511-a060-4d30-af62-872ba84b84c3-kube-api-access-hf9h9\") pod \"manila-operator-controller-manager-5d499bf58b-5hvj7\" (UID: \"021e1511-a060-4d30-af62-872ba84b84c3\") " pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.269784 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.290442 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2s825"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.292203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.294505 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-w5r7f" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.297311 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2s825"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.310440 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.313903 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.317773 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-bc5rz" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.324125 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.325578 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368524 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5vdp\" (UniqueName: \"kubernetes.io/projected/7ac95117-0ba0-4c51-afbf-320322b8ebfc-kube-api-access-v5vdp\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w8ng\" (UniqueName: \"kubernetes.io/projected/f9b85bd6-21c3-457e-81c0-b87c56be9f3e-kube-api-access-9w8ng\") pod \"placement-operator-controller-manager-57988cc5b5-ffw7r\" (UID: \"f9b85bd6-21c3-457e-81c0-b87c56be9f3e\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368664 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-467qp\" (UniqueName: \"kubernetes.io/projected/7bead1ee-7a42-45c9-9205-420ae85002f7-kube-api-access-467qp\") pod \"neutron-operator-controller-manager-6fdcddb789-q97hn\" (UID: \"7bead1ee-7a42-45c9-9205-420ae85002f7\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368681 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zclns\" (UniqueName: \"kubernetes.io/projected/d22fe1a8-9183-4ca4-ae47-60d3e77f2339-kube-api-access-zclns\") pod \"octavia-operator-controller-manager-64cdc6ff96-plpx4\" (UID: \"d22fe1a8-9183-4ca4-ae47-60d3e77f2339\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368820 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb47r\" (UniqueName: \"kubernetes.io/projected/85c90cee-6304-4ab9-957c-0bd91411403f-kube-api-access-vb47r\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-6jrgg\" (UID: \"85c90cee-6304-4ab9-957c-0bd91411403f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368908 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-299bx\" (UniqueName: \"kubernetes.io/projected/d8ad3c0b-931e-444d-9fcf-23dfbe77bedc-kube-api-access-299bx\") pod \"nova-operator-controller-manager-79556f57fc-ftsgz\" (UID: \"d8ad3c0b-931e-444d-9fcf-23dfbe77bedc\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.368925 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.369021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms2vw\" (UniqueName: \"kubernetes.io/projected/5e867f5f-7847-4469-95cb-605042211f56-kube-api-access-ms2vw\") pod \"ovn-operator-controller-manager-56897c768d-ksct6\" (UID: \"5e867f5f-7847-4469-95cb-605042211f56\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.374365 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.382192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.443863 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.457348 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.462112 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-zl2mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.467641 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469739 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w8ng\" (UniqueName: \"kubernetes.io/projected/f9b85bd6-21c3-457e-81c0-b87c56be9f3e-kube-api-access-9w8ng\") pod \"placement-operator-controller-manager-57988cc5b5-ffw7r\" (UID: \"f9b85bd6-21c3-457e-81c0-b87c56be9f3e\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469777 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-467qp\" (UniqueName: \"kubernetes.io/projected/7bead1ee-7a42-45c9-9205-420ae85002f7-kube-api-access-467qp\") pod \"neutron-operator-controller-manager-6fdcddb789-q97hn\" (UID: \"7bead1ee-7a42-45c9-9205-420ae85002f7\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469794 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zclns\" (UniqueName: \"kubernetes.io/projected/d22fe1a8-9183-4ca4-ae47-60d3e77f2339-kube-api-access-zclns\") pod \"octavia-operator-controller-manager-64cdc6ff96-plpx4\" (UID: \"d22fe1a8-9183-4ca4-ae47-60d3e77f2339\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cth7r\" (UniqueName: \"kubernetes.io/projected/04b644f3-2f69-4cb5-8ade-04f4be267255-kube-api-access-cth7r\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fkb27\" (UID: \"04b644f3-2f69-4cb5-8ade-04f4be267255\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4csx\" (UniqueName: \"kubernetes.io/projected/523ab37c-4bb4-4d00-8134-ad63c2833907-kube-api-access-s4csx\") pod \"test-operator-controller-manager-5cd6c7f4c8-8sjts\" (UID: \"523ab37c-4bb4-4d00-8134-ad63c2833907\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469886 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb47r\" (UniqueName: \"kubernetes.io/projected/85c90cee-6304-4ab9-957c-0bd91411403f-kube-api-access-vb47r\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-6jrgg\" (UID: \"85c90cee-6304-4ab9-957c-0bd91411403f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469908 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-299bx\" (UniqueName: \"kubernetes.io/projected/d8ad3c0b-931e-444d-9fcf-23dfbe77bedc-kube-api-access-299bx\") pod \"nova-operator-controller-manager-79556f57fc-ftsgz\" (UID: \"d8ad3c0b-931e-444d-9fcf-23dfbe77bedc\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469946 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms2vw\" (UniqueName: \"kubernetes.io/projected/5e867f5f-7847-4469-95cb-605042211f56-kube-api-access-ms2vw\") pod \"ovn-operator-controller-manager-56897c768d-ksct6\" (UID: \"5e867f5f-7847-4469-95cb-605042211f56\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469963 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dwng\" (UniqueName: \"kubernetes.io/projected/e72bf30c-6c74-46a6-b215-ac5b1f33c144-kube-api-access-7dwng\") pod \"swift-operator-controller-manager-d77b94747-2s825\" (UID: \"e72bf30c-6c74-46a6-b215-ac5b1f33c144\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.469993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5vdp\" (UniqueName: \"kubernetes.io/projected/7ac95117-0ba0-4c51-afbf-320322b8ebfc-kube-api-access-v5vdp\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.470430 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.470476 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert podName:7ac95117-0ba0-4c51-afbf-320322b8ebfc nodeName:}" failed. No retries permitted until 2025-11-26 06:38:37.970460232 +0000 UTC m=+801.331764274 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" (UID: "7ac95117-0ba0-4c51-afbf-320322b8ebfc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.499480 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.502750 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w8ng\" (UniqueName: \"kubernetes.io/projected/f9b85bd6-21c3-457e-81c0-b87c56be9f3e-kube-api-access-9w8ng\") pod \"placement-operator-controller-manager-57988cc5b5-ffw7r\" (UID: \"f9b85bd6-21c3-457e-81c0-b87c56be9f3e\") " pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.506428 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-299bx\" (UniqueName: \"kubernetes.io/projected/d8ad3c0b-931e-444d-9fcf-23dfbe77bedc-kube-api-access-299bx\") pod \"nova-operator-controller-manager-79556f57fc-ftsgz\" (UID: \"d8ad3c0b-931e-444d-9fcf-23dfbe77bedc\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.506957 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5vdp\" (UniqueName: \"kubernetes.io/projected/7ac95117-0ba0-4c51-afbf-320322b8ebfc-kube-api-access-v5vdp\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.509429 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb47r\" (UniqueName: \"kubernetes.io/projected/85c90cee-6304-4ab9-957c-0bd91411403f-kube-api-access-vb47r\") pod \"mariadb-operator-controller-manager-66f4dd4bc7-6jrgg\" (UID: \"85c90cee-6304-4ab9-957c-0bd91411403f\") " pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.511969 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms2vw\" (UniqueName: \"kubernetes.io/projected/5e867f5f-7847-4469-95cb-605042211f56-kube-api-access-ms2vw\") pod \"ovn-operator-controller-manager-56897c768d-ksct6\" (UID: \"5e867f5f-7847-4469-95cb-605042211f56\") " pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.514556 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zclns\" (UniqueName: \"kubernetes.io/projected/d22fe1a8-9183-4ca4-ae47-60d3e77f2339-kube-api-access-zclns\") pod \"octavia-operator-controller-manager-64cdc6ff96-plpx4\" (UID: \"d22fe1a8-9183-4ca4-ae47-60d3e77f2339\") " pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.517995 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-467qp\" (UniqueName: \"kubernetes.io/projected/7bead1ee-7a42-45c9-9205-420ae85002f7-kube-api-access-467qp\") pod \"neutron-operator-controller-manager-6fdcddb789-q97hn\" (UID: \"7bead1ee-7a42-45c9-9205-420ae85002f7\") " pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.518274 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.540158 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.550808 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.564790 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.566856 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.567294 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.569309 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-swbmr" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.571501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cth7r\" (UniqueName: \"kubernetes.io/projected/04b644f3-2f69-4cb5-8ade-04f4be267255-kube-api-access-cth7r\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fkb27\" (UID: \"04b644f3-2f69-4cb5-8ade-04f4be267255\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.571551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4csx\" (UniqueName: \"kubernetes.io/projected/523ab37c-4bb4-4d00-8134-ad63c2833907-kube-api-access-s4csx\") pod \"test-operator-controller-manager-5cd6c7f4c8-8sjts\" (UID: \"523ab37c-4bb4-4d00-8134-ad63c2833907\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.571605 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dwng\" (UniqueName: \"kubernetes.io/projected/e72bf30c-6c74-46a6-b215-ac5b1f33c144-kube-api-access-7dwng\") pod \"swift-operator-controller-manager-d77b94747-2s825\" (UID: \"e72bf30c-6c74-46a6-b215-ac5b1f33c144\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.571660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wgh8\" (UniqueName: \"kubernetes.io/projected/0fa640fb-a493-4f84-bfcf-a3824ddc061f-kube-api-access-5wgh8\") pod \"watcher-operator-controller-manager-656dcb59d4-65qdq\" (UID: \"0fa640fb-a493-4f84-bfcf-a3824ddc061f\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.571928 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.579504 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.580682 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.586304 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.587456 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.587624 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.587823 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-6frqr" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.589184 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.600608 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dwng\" (UniqueName: \"kubernetes.io/projected/e72bf30c-6c74-46a6-b215-ac5b1f33c144-kube-api-access-7dwng\") pod \"swift-operator-controller-manager-d77b94747-2s825\" (UID: \"e72bf30c-6c74-46a6-b215-ac5b1f33c144\") " pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.602610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4csx\" (UniqueName: \"kubernetes.io/projected/523ab37c-4bb4-4d00-8134-ad63c2833907-kube-api-access-s4csx\") pod \"test-operator-controller-manager-5cd6c7f4c8-8sjts\" (UID: \"523ab37c-4bb4-4d00-8134-ad63c2833907\") " pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.606039 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.613109 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.613934 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cth7r\" (UniqueName: \"kubernetes.io/projected/04b644f3-2f69-4cb5-8ade-04f4be267255-kube-api-access-cth7r\") pod \"telemetry-operator-controller-manager-76cc84c6bb-fkb27\" (UID: \"04b644f3-2f69-4cb5-8ade-04f4be267255\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.615454 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-g4dll" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.616817 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.619935 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.637170 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.672583 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wgh8\" (UniqueName: \"kubernetes.io/projected/0fa640fb-a493-4f84-bfcf-a3824ddc061f-kube-api-access-5wgh8\") pod \"watcher-operator-controller-manager-656dcb59d4-65qdq\" (UID: \"0fa640fb-a493-4f84-bfcf-a3824ddc061f\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.672634 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb2lw\" (UniqueName: \"kubernetes.io/projected/2e76307b-7079-41d7-a66e-35f03cc1a1bd-kube-api-access-rb2lw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-r7dqc\" (UID: \"2e76307b-7079-41d7-a66e-35f03cc1a1bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.672686 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.672739 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp5k5\" (UniqueName: \"kubernetes.io/projected/cf2135a7-5206-498c-9391-379fae5f4bfa-kube-api-access-dp5k5\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.672775 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.672799 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.672866 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.672925 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert podName:8483a9f8-d70d-497d-ae02-c6b426a49306 nodeName:}" failed. No retries permitted until 2025-11-26 06:38:38.672905318 +0000 UTC m=+802.034209270 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert") pod "infra-operator-controller-manager-57548d458d-xrzlh" (UID: "8483a9f8-d70d-497d-ae02-c6b426a49306") : secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.689782 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.692256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wgh8\" (UniqueName: \"kubernetes.io/projected/0fa640fb-a493-4f84-bfcf-a3824ddc061f-kube-api-access-5wgh8\") pod \"watcher-operator-controller-manager-656dcb59d4-65qdq\" (UID: \"0fa640fb-a493-4f84-bfcf-a3824ddc061f\") " pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.702976 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-955677c94-dghfx"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.706096 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.776649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp5k5\" (UniqueName: \"kubernetes.io/projected/cf2135a7-5206-498c-9391-379fae5f4bfa-kube-api-access-dp5k5\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.776729 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.776769 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.776815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb2lw\" (UniqueName: \"kubernetes.io/projected/2e76307b-7079-41d7-a66e-35f03cc1a1bd-kube-api-access-rb2lw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-r7dqc\" (UID: \"2e76307b-7079-41d7-a66e-35f03cc1a1bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.777101 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.777159 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:38.277137435 +0000 UTC m=+801.638441387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "metrics-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.777302 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.777369 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:38.277349771 +0000 UTC m=+801.638653793 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.795256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp5k5\" (UniqueName: \"kubernetes.io/projected/cf2135a7-5206-498c-9391-379fae5f4bfa-kube-api-access-dp5k5\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.798390 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb2lw\" (UniqueName: \"kubernetes.io/projected/2e76307b-7079-41d7-a66e-35f03cc1a1bd-kube-api-access-rb2lw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-r7dqc\" (UID: \"2e76307b-7079-41d7-a66e-35f03cc1a1bd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.801013 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.819229 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.826575 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.833182 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9"] Nov 26 06:38:37 crc kubenswrapper[4775]: W1126 06:38:37.841617 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod997d00b8_4cc3_45e9_8af2_2d573eef844a.slice/crio-218fc1819f876b1a2461952e6d4289838984413a943484d96c4e6f78f2782447 WatchSource:0}: Error finding container 218fc1819f876b1a2461952e6d4289838984413a943484d96c4e6f78f2782447: Status 404 returned error can't find the container with id 218fc1819f876b1a2461952e6d4289838984413a943484d96c4e6f78f2782447 Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.884482 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.945780 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.980951 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d"] Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.983186 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.983847 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:37 crc kubenswrapper[4775]: E1126 06:38:37.983929 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert podName:7ac95117-0ba0-4c51-afbf-320322b8ebfc nodeName:}" failed. No retries permitted until 2025-11-26 06:38:38.983914475 +0000 UTC m=+802.345218427 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" (UID: "7ac95117-0ba0-4c51-afbf-320322b8ebfc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.986762 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v"] Nov 26 06:38:37 crc kubenswrapper[4775]: I1126 06:38:37.993306 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.166301 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.261563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.267709 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7"] Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.268973 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod021e1511_a060_4d30_af62_872ba84b84c3.slice/crio-e1cd006980dcc9730070d8e599135e6cbd68094a90278c6ebf438f9cef0f1805 WatchSource:0}: Error finding container e1cd006980dcc9730070d8e599135e6cbd68094a90278c6ebf438f9cef0f1805: Status 404 returned error can't find the container with id e1cd006980dcc9730070d8e599135e6cbd68094a90278c6ebf438f9cef0f1805 Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.269145 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" event={"ID":"ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7","Type":"ContainerStarted","Data":"bffd24db9ef40cbbfef4e677689121077256fe2d81def4a495a17e04ef58dfc0"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.271329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" event={"ID":"cc351f29-006f-4ee4-80cb-5860f4f824ff","Type":"ContainerStarted","Data":"f2320ebbb5ab2a3e2cb0d6ee410da93b5efd853c346dd4ad65af04def8ff9b88"} Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.271471 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85c90cee_6304_4ab9_957c_0bd91411403f.slice/crio-c07d7bbb15ff753e0113482c99a3b679bc0f212656ef9c0d7ae817b56feaecf7 WatchSource:0}: Error finding container c07d7bbb15ff753e0113482c99a3b679bc0f212656ef9c0d7ae817b56feaecf7: Status 404 returned error can't find the container with id c07d7bbb15ff753e0113482c99a3b679bc0f212656ef9c0d7ae817b56feaecf7 Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.272707 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" event={"ID":"2158b3e8-3c08-4e7c-9b28-cd32534135a1","Type":"ContainerStarted","Data":"0defeaaa34e81d35ca442ccba8574e15fcf05128581efb1f2e958e89556ccb88"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.277473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" event={"ID":"153b27e4-d948-4ff6-9d6a-bfbcebd17cad","Type":"ContainerStarted","Data":"af8940dc0f4ea0f935164642f05f3e0e048361f69873f70521dab21a7fd76d67"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.283762 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" event={"ID":"4f5effa7-3458-4978-8382-0f5b4a17105a","Type":"ContainerStarted","Data":"39a5a5d2c02b3a92a4c0589e851f30abeccc707c7b2f75a8fd94061e3e9388f6"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.284599 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" event={"ID":"faa5a2e8-ca2b-42e1-ac9a-3580111707ac","Type":"ContainerStarted","Data":"8a818be1cc71fa1d128bca0fe60fd760ddc000244e91fc420e1034937a976258"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.287874 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" event={"ID":"1a1d393e-2d6c-425c-9e6f-068d4c8ee090","Type":"ContainerStarted","Data":"7e52ad1a26194f1048505d039498b91df73eb935ebba4e2faed9dcc738f67daf"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.288947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" event={"ID":"997d00b8-4cc3-45e9-8af2-2d573eef844a","Type":"ContainerStarted","Data":"218fc1819f876b1a2461952e6d4289838984413a943484d96c4e6f78f2782447"} Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.289650 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.299865 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.300069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.300322 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.300426 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:39.300410115 +0000 UTC m=+802.661714067 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "webhook-server-cert" not found Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.300883 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.300980 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:39.30095814 +0000 UTC m=+802.662262112 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "metrics-server-cert" not found Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.386342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4"] Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.389568 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8ad3c0b_931e_444d_9fcf_23dfbe77bedc.slice/crio-d677833f30ee4f14a9ccd53c4db5323c6f616d04625a6caf3cb50d98fa6f1c93 WatchSource:0}: Error finding container d677833f30ee4f14a9ccd53c4db5323c6f616d04625a6caf3cb50d98fa6f1c93: Status 404 returned error can't find the container with id d677833f30ee4f14a9ccd53c4db5323c6f616d04625a6caf3cb50d98fa6f1c93 Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.389999 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd22fe1a8_9183_4ca4_ae47_60d3e77f2339.slice/crio-60ae936fadade8f2e4bfcb47d5a8afca51287e51684aad9bdd84e792cbff086e WatchSource:0}: Error finding container 60ae936fadade8f2e4bfcb47d5a8afca51287e51684aad9bdd84e792cbff086e: Status 404 returned error can't find the container with id 60ae936fadade8f2e4bfcb47d5a8afca51287e51684aad9bdd84e792cbff086e Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.392161 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz"] Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.395952 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zclns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-plpx4_openstack-operators(d22fe1a8-9183-4ca4-ae47-60d3e77f2339): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.398676 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zclns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-64cdc6ff96-plpx4_openstack-operators(d22fe1a8-9183-4ca4-ae47-60d3e77f2339): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.405163 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" podUID="d22fe1a8-9183-4ca4-ae47-60d3e77f2339" Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.457719 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d77b94747-2s825"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.462888 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r"] Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.464115 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9b85bd6_21c3_457e_81c0_b87c56be9f3e.slice/crio-79f71888ce4372bfb5d300437f52157f3b146d827e72578ec543976c5e3f2c09 WatchSource:0}: Error finding container 79f71888ce4372bfb5d300437f52157f3b146d827e72578ec543976c5e3f2c09: Status 404 returned error can't find the container with id 79f71888ce4372bfb5d300437f52157f3b146d827e72578ec543976c5e3f2c09 Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.473577 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode72bf30c_6c74_46a6_b215_ac5b1f33c144.slice/crio-8f5f80546e6345ce64bfe4cb119aadf939f0a7967fb70f6d0313ac72bc8a00b8 WatchSource:0}: Error finding container 8f5f80546e6345ce64bfe4cb119aadf939f0a7967fb70f6d0313ac72bc8a00b8: Status 404 returned error can't find the container with id 8f5f80546e6345ce64bfe4cb119aadf939f0a7967fb70f6d0313ac72bc8a00b8 Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.476649 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7dwng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-2s825_openstack-operators(e72bf30c-6c74-46a6-b215-ac5b1f33c144): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.478876 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7dwng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-d77b94747-2s825_openstack-operators(e72bf30c-6c74-46a6-b215-ac5b1f33c144): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.480956 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" podUID="e72bf30c-6c74-46a6-b215-ac5b1f33c144" Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.552076 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.568825 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.574983 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6"] Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.582292 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq"] Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.585337 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cth7r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-fkb27_openstack-operators(04b644f3-2f69-4cb5-8ade-04f4be267255): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.585381 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rb2lw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-r7dqc_openstack-operators(2e76307b-7079-41d7-a66e-35f03cc1a1bd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.585883 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fa640fb_a493_4f84_bfcf_a3824ddc061f.slice/crio-fb128318add15e60bf7956528262c8a418d6891ca9fc898ba3657b104a29da15 WatchSource:0}: Error finding container fb128318add15e60bf7956528262c8a418d6891ca9fc898ba3657b104a29da15: Status 404 returned error can't find the container with id fb128318add15e60bf7956528262c8a418d6891ca9fc898ba3657b104a29da15 Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.586016 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27"] Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.586602 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" podUID="2e76307b-7079-41d7-a66e-35f03cc1a1bd" Nov 26 06:38:38 crc kubenswrapper[4775]: W1126 06:38:38.587705 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e867f5f_7847_4469_95cb_605042211f56.slice/crio-c8d2fbcb4eda7b3ddf46edf7c62ba3b829754a63ad689fa9d123d25994bd295f WatchSource:0}: Error finding container c8d2fbcb4eda7b3ddf46edf7c62ba3b829754a63ad689fa9d123d25994bd295f: Status 404 returned error can't find the container with id c8d2fbcb4eda7b3ddf46edf7c62ba3b829754a63ad689fa9d123d25994bd295f Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.587924 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cth7r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-fkb27_openstack-operators(04b644f3-2f69-4cb5-8ade-04f4be267255): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.589039 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" podUID="04b644f3-2f69-4cb5-8ade-04f4be267255" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.589909 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5wgh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-65qdq_openstack-operators(0fa640fb-a493-4f84-bfcf-a3824ddc061f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.591018 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ms2vw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-ksct6_openstack-operators(5e867f5f-7847-4469-95cb-605042211f56): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.591025 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s4csx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-8sjts_openstack-operators(523ab37c-4bb4-4d00-8134-ad63c2833907): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.591595 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5wgh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-656dcb59d4-65qdq_openstack-operators(0fa640fb-a493-4f84-bfcf-a3824ddc061f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.592689 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" podUID="0fa640fb-a493-4f84-bfcf-a3824ddc061f" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.593470 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ms2vw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-56897c768d-ksct6_openstack-operators(5e867f5f-7847-4469-95cb-605042211f56): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.594011 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s4csx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-8sjts_openstack-operators(523ab37c-4bb4-4d00-8134-ad63c2833907): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.595211 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podUID="523ab37c-4bb4-4d00-8134-ad63c2833907" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.595242 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" podUID="5e867f5f-7847-4469-95cb-605042211f56" Nov 26 06:38:38 crc kubenswrapper[4775]: I1126 06:38:38.705626 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.705779 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:38 crc kubenswrapper[4775]: E1126 06:38:38.705833 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert podName:8483a9f8-d70d-497d-ae02-c6b426a49306 nodeName:}" failed. No retries permitted until 2025-11-26 06:38:40.70581919 +0000 UTC m=+804.067123142 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert") pod "infra-operator-controller-manager-57548d458d-xrzlh" (UID: "8483a9f8-d70d-497d-ae02-c6b426a49306") : secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.009255 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.009397 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.009454 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert podName:7ac95117-0ba0-4c51-afbf-320322b8ebfc nodeName:}" failed. No retries permitted until 2025-11-26 06:38:41.009440132 +0000 UTC m=+804.370744084 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" (UID: "7ac95117-0ba0-4c51-afbf-320322b8ebfc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.296977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" event={"ID":"85c90cee-6304-4ab9-957c-0bd91411403f","Type":"ContainerStarted","Data":"c07d7bbb15ff753e0113482c99a3b679bc0f212656ef9c0d7ae817b56feaecf7"} Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.299296 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" event={"ID":"d8ad3c0b-931e-444d-9fcf-23dfbe77bedc","Type":"ContainerStarted","Data":"d677833f30ee4f14a9ccd53c4db5323c6f616d04625a6caf3cb50d98fa6f1c93"} Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.302411 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" event={"ID":"7bead1ee-7a42-45c9-9205-420ae85002f7","Type":"ContainerStarted","Data":"9b235089b58e4189c12c97e59191da583ba95c0b522be1efd4ec9ee6bcc405e5"} Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.303880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" event={"ID":"5e867f5f-7847-4469-95cb-605042211f56","Type":"ContainerStarted","Data":"c8d2fbcb4eda7b3ddf46edf7c62ba3b829754a63ad689fa9d123d25994bd295f"} Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.305788 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" podUID="5e867f5f-7847-4469-95cb-605042211f56" Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.307426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" event={"ID":"f9b85bd6-21c3-457e-81c0-b87c56be9f3e","Type":"ContainerStarted","Data":"79f71888ce4372bfb5d300437f52157f3b146d827e72578ec543976c5e3f2c09"} Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.310367 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" event={"ID":"04b644f3-2f69-4cb5-8ade-04f4be267255","Type":"ContainerStarted","Data":"9410766d3b5fff7d2cabfdbc738f20c957eb16d54f010b66897da3561bccabde"} Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.312189 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" event={"ID":"e72bf30c-6c74-46a6-b215-ac5b1f33c144","Type":"ContainerStarted","Data":"8f5f80546e6345ce64bfe4cb119aadf939f0a7967fb70f6d0313ac72bc8a00b8"} Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.312940 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" podUID="04b644f3-2f69-4cb5-8ade-04f4be267255" Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.313468 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.313511 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.313623 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.313638 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.313676 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:41.31366149 +0000 UTC m=+804.674965442 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "metrics-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.313692 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:41.313684471 +0000 UTC m=+804.674988423 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "webhook-server-cert" not found Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.314901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" event={"ID":"021e1511-a060-4d30-af62-872ba84b84c3","Type":"ContainerStarted","Data":"e1cd006980dcc9730070d8e599135e6cbd68094a90278c6ebf438f9cef0f1805"} Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.316202 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" podUID="e72bf30c-6c74-46a6-b215-ac5b1f33c144" Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.317128 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" event={"ID":"523ab37c-4bb4-4d00-8134-ad63c2833907","Type":"ContainerStarted","Data":"82b4710a953237931e16704b608c661342b2d9e11a629d5b87cb725c550a0915"} Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.320443 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podUID="523ab37c-4bb4-4d00-8134-ad63c2833907" Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.321630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" event={"ID":"2e76307b-7079-41d7-a66e-35f03cc1a1bd","Type":"ContainerStarted","Data":"ae570437235fe7b03f95c27ba0737ee2a95201ac09b1d87cdf7cbc177f9f0147"} Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.323101 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" podUID="2e76307b-7079-41d7-a66e-35f03cc1a1bd" Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.325238 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" event={"ID":"0fa640fb-a493-4f84-bfcf-a3824ddc061f","Type":"ContainerStarted","Data":"fb128318add15e60bf7956528262c8a418d6891ca9fc898ba3657b104a29da15"} Nov 26 06:38:39 crc kubenswrapper[4775]: I1126 06:38:39.326609 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" event={"ID":"d22fe1a8-9183-4ca4-ae47-60d3e77f2339","Type":"ContainerStarted","Data":"60ae936fadade8f2e4bfcb47d5a8afca51287e51684aad9bdd84e792cbff086e"} Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.327445 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" podUID="0fa640fb-a493-4f84-bfcf-a3824ddc061f" Nov 26 06:38:39 crc kubenswrapper[4775]: E1126 06:38:39.330029 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" podUID="d22fe1a8-9183-4ca4-ae47-60d3e77f2339" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.338673 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podUID="523ab37c-4bb4-4d00-8134-ad63c2833907" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.339023 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" podUID="2e76307b-7079-41d7-a66e-35f03cc1a1bd" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.339064 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72236301580ff9080f7e311b832d7ba66666a9afeda51f969745229624ff26e4\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" podUID="e72bf30c-6c74-46a6-b215-ac5b1f33c144" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.339276 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6bed55b172b9ee8ccc3952cbfc543d8bd44e2690f6db94348a754152fd78f4cf\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" podUID="0fa640fb-a493-4f84-bfcf-a3824ddc061f" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.340528 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:ddc8a82f05930db8ee7a8d6d189b5a66373060656e4baf71ac302f89c477da4c\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" podUID="d22fe1a8-9183-4ca4-ae47-60d3e77f2339" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.340575 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" podUID="04b644f3-2f69-4cb5-8ade-04f4be267255" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.341529 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bbb543d2d67c73e5df5d6357c3251363eb34a99575c5bf10416edd45dbdae2f6\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" podUID="5e867f5f-7847-4469-95cb-605042211f56" Nov 26 06:38:40 crc kubenswrapper[4775]: I1126 06:38:40.736845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.737042 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:40 crc kubenswrapper[4775]: E1126 06:38:40.737118 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert podName:8483a9f8-d70d-497d-ae02-c6b426a49306 nodeName:}" failed. No retries permitted until 2025-11-26 06:38:44.737099666 +0000 UTC m=+808.098403618 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert") pod "infra-operator-controller-manager-57548d458d-xrzlh" (UID: "8483a9f8-d70d-497d-ae02-c6b426a49306") : secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:41 crc kubenswrapper[4775]: I1126 06:38:41.039997 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:41 crc kubenswrapper[4775]: E1126 06:38:41.040212 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:41 crc kubenswrapper[4775]: E1126 06:38:41.040257 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert podName:7ac95117-0ba0-4c51-afbf-320322b8ebfc nodeName:}" failed. No retries permitted until 2025-11-26 06:38:45.040243996 +0000 UTC m=+808.401547948 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" (UID: "7ac95117-0ba0-4c51-afbf-320322b8ebfc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:41 crc kubenswrapper[4775]: I1126 06:38:41.343743 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:41 crc kubenswrapper[4775]: I1126 06:38:41.344123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:41 crc kubenswrapper[4775]: E1126 06:38:41.343922 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 06:38:41 crc kubenswrapper[4775]: E1126 06:38:41.344352 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:45.344335571 +0000 UTC m=+808.705639533 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "metrics-server-cert" not found Nov 26 06:38:41 crc kubenswrapper[4775]: E1126 06:38:41.344295 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 06:38:41 crc kubenswrapper[4775]: E1126 06:38:41.344827 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:45.344813024 +0000 UTC m=+808.706116986 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "webhook-server-cert" not found Nov 26 06:38:44 crc kubenswrapper[4775]: I1126 06:38:44.794225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:44 crc kubenswrapper[4775]: E1126 06:38:44.794442 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:44 crc kubenswrapper[4775]: E1126 06:38:44.794508 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert podName:8483a9f8-d70d-497d-ae02-c6b426a49306 nodeName:}" failed. No retries permitted until 2025-11-26 06:38:52.794486914 +0000 UTC m=+816.155790866 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert") pod "infra-operator-controller-manager-57548d458d-xrzlh" (UID: "8483a9f8-d70d-497d-ae02-c6b426a49306") : secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:45 crc kubenswrapper[4775]: I1126 06:38:45.099381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:45 crc kubenswrapper[4775]: E1126 06:38:45.099464 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:45 crc kubenswrapper[4775]: E1126 06:38:45.099702 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert podName:7ac95117-0ba0-4c51-afbf-320322b8ebfc nodeName:}" failed. No retries permitted until 2025-11-26 06:38:53.099659637 +0000 UTC m=+816.460963589 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert") pod "openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" (UID: "7ac95117-0ba0-4c51-afbf-320322b8ebfc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 26 06:38:45 crc kubenswrapper[4775]: I1126 06:38:45.403898 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:45 crc kubenswrapper[4775]: I1126 06:38:45.403985 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:45 crc kubenswrapper[4775]: E1126 06:38:45.404141 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 26 06:38:45 crc kubenswrapper[4775]: E1126 06:38:45.404197 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:53.404179353 +0000 UTC m=+816.765483305 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "webhook-server-cert" not found Nov 26 06:38:45 crc kubenswrapper[4775]: E1126 06:38:45.404601 4775 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 26 06:38:45 crc kubenswrapper[4775]: E1126 06:38:45.404734 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs podName:cf2135a7-5206-498c-9391-379fae5f4bfa nodeName:}" failed. No retries permitted until 2025-11-26 06:38:53.404684696 +0000 UTC m=+816.765988688 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs") pod "openstack-operator-controller-manager-646778448f-7d5mc" (UID: "cf2135a7-5206-498c-9391-379fae5f4bfa") : secret "metrics-server-cert" not found Nov 26 06:38:49 crc kubenswrapper[4775]: E1126 06:38:49.808337 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:d65dbfc956e9cf376f3c48fc3a0942cb7306b5164f898c40d1efca106df81db7" Nov 26 06:38:49 crc kubenswrapper[4775]: E1126 06:38:49.809392 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:d65dbfc956e9cf376f3c48fc3a0942cb7306b5164f898c40d1efca106df81db7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t98bv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-67cb4dc6d4-gdk4v_openstack-operators(2158b3e8-3c08-4e7c-9b28-cd32534135a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:38:50 crc kubenswrapper[4775]: E1126 06:38:50.470072 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423" Nov 26 06:38:50 crc kubenswrapper[4775]: E1126 06:38:50.470260 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9w8ng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-57988cc5b5-ffw7r_openstack-operators(f9b85bd6-21c3-457e-81c0-b87c56be9f3e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:38:51 crc kubenswrapper[4775]: E1126 06:38:51.010381 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vb47r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-66f4dd4bc7-6jrgg_openstack-operators(85c90cee-6304-4ab9-957c-0bd91411403f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:51 crc kubenswrapper[4775]: E1126 06:38:51.013032 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" podUID="85c90cee-6304-4ab9-957c-0bd91411403f" Nov 26 06:38:51 crc kubenswrapper[4775]: E1126 06:38:51.337394 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s4csx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-8sjts_openstack-operators(523ab37c-4bb4-4d00-8134-ad63c2833907): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:51 crc kubenswrapper[4775]: E1126 06:38:51.339442 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s4csx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cd6c7f4c8-8sjts_openstack-operators(523ab37c-4bb4-4d00-8134-ad63c2833907): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 26 06:38:51 crc kubenswrapper[4775]: E1126 06:38:51.340580 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podUID="523ab37c-4bb4-4d00-8134-ad63c2833907" Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.433296 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" event={"ID":"997d00b8-4cc3-45e9-8af2-2d573eef844a","Type":"ContainerStarted","Data":"2f2f315bc8a7e05ed22f82cef79592b653eb5fecee8093fda53826d3e4f7e814"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.435738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" event={"ID":"ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7","Type":"ContainerStarted","Data":"566a112c99148d2de70dbf6d856252d44c9bc355235e94afb25fee443c6ff408"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.437192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" event={"ID":"021e1511-a060-4d30-af62-872ba84b84c3","Type":"ContainerStarted","Data":"73f6b2abc81ee862e2d3b78b769427587c03d7c4d94a846ea2699162fc406033"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.438461 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" event={"ID":"153b27e4-d948-4ff6-9d6a-bfbcebd17cad","Type":"ContainerStarted","Data":"c26809574b612cf9215285e374d0c5474ae684de93ebe93038fd63a702461f9f"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.446355 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" event={"ID":"4f5effa7-3458-4978-8382-0f5b4a17105a","Type":"ContainerStarted","Data":"aad15f0dad51b6a8cf6a18cdfc2a3223264b24bffc3bdead9f176a36cb1b9251"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.447667 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" event={"ID":"1a1d393e-2d6c-425c-9e6f-068d4c8ee090","Type":"ContainerStarted","Data":"04b4d00afc28f8c8d7a5bbb47ea98c29a85feadf226e26bcd8d464fc329aaa3f"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.450516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" event={"ID":"85c90cee-6304-4ab9-957c-0bd91411403f","Type":"ContainerStarted","Data":"ccd4e731e6a779364eede3e11e632eee506c0c9f7725146a760e5a5227eae2fd"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.451175 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:38:51 crc kubenswrapper[4775]: E1126 06:38:51.451887 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" podUID="85c90cee-6304-4ab9-957c-0bd91411403f" Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.453974 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" event={"ID":"d8ad3c0b-931e-444d-9fcf-23dfbe77bedc","Type":"ContainerStarted","Data":"ec89de56840af1458476a8cdfaa92722f5858916ec4224f2d21b1eaf25316ba8"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.470491 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" event={"ID":"7bead1ee-7a42-45c9-9205-420ae85002f7","Type":"ContainerStarted","Data":"cf63dbcb2a8239856ec855bf4e8c8c6d034ca74ce17b869781c023418a94b147"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.481208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" event={"ID":"cc351f29-006f-4ee4-80cb-5860f4f824ff","Type":"ContainerStarted","Data":"293f26141351ca0b3f39954d6740710c4800d9a0cbcecaae61760423367db516"} Nov 26 06:38:51 crc kubenswrapper[4775]: I1126 06:38:51.488398 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" event={"ID":"faa5a2e8-ca2b-42e1-ac9a-3580111707ac","Type":"ContainerStarted","Data":"1122d783a2532733db6bbfa03a73e6a6d8ffe7c62a91302b76cb58ed367763a1"} Nov 26 06:38:52 crc kubenswrapper[4775]: E1126 06:38:52.501976 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" podUID="85c90cee-6304-4ab9-957c-0bd91411403f" Nov 26 06:38:52 crc kubenswrapper[4775]: I1126 06:38:52.839950 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:38:52 crc kubenswrapper[4775]: E1126 06:38:52.840197 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:52 crc kubenswrapper[4775]: E1126 06:38:52.840252 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert podName:8483a9f8-d70d-497d-ae02-c6b426a49306 nodeName:}" failed. No retries permitted until 2025-11-26 06:39:08.840236177 +0000 UTC m=+832.201540129 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert") pod "infra-operator-controller-manager-57548d458d-xrzlh" (UID: "8483a9f8-d70d-497d-ae02-c6b426a49306") : secret "infra-operator-webhook-server-cert" not found Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.143779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.149337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ac95117-0ba0-4c51-afbf-320322b8ebfc-cert\") pod \"openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv\" (UID: \"7ac95117-0ba0-4c51-afbf-320322b8ebfc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.199691 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.447348 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.447398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.453822 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-webhook-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.457447 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cf2135a7-5206-498c-9391-379fae5f4bfa-metrics-certs\") pod \"openstack-operator-controller-manager-646778448f-7d5mc\" (UID: \"cf2135a7-5206-498c-9391-379fae5f4bfa\") " pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:53 crc kubenswrapper[4775]: I1126 06:38:53.500140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:54 crc kubenswrapper[4775]: I1126 06:38:54.695100 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc"] Nov 26 06:38:54 crc kubenswrapper[4775]: W1126 06:38:54.728514 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf2135a7_5206_498c_9391_379fae5f4bfa.slice/crio-7e8043ab0c7d11cec7bd521c252b93f4021a4e188d28d5bfa30478986d19a867 WatchSource:0}: Error finding container 7e8043ab0c7d11cec7bd521c252b93f4021a4e188d28d5bfa30478986d19a867: Status 404 returned error can't find the container with id 7e8043ab0c7d11cec7bd521c252b93f4021a4e188d28d5bfa30478986d19a867 Nov 26 06:38:54 crc kubenswrapper[4775]: I1126 06:38:54.838383 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv"] Nov 26 06:38:55 crc kubenswrapper[4775]: E1126 06:38:55.106865 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" podUID="2158b3e8-3c08-4e7c-9b28-cd32534135a1" Nov 26 06:38:55 crc kubenswrapper[4775]: E1126 06:38:55.168096 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" podUID="f9b85bd6-21c3-457e-81c0-b87c56be9f3e" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.526507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" event={"ID":"f9b85bd6-21c3-457e-81c0-b87c56be9f3e","Type":"ContainerStarted","Data":"eecd5e7ee181145cbe3d983be380fce7977cfc04fe6a0cfa38d659666c5296f9"} Nov 26 06:38:55 crc kubenswrapper[4775]: E1126 06:38:55.528597 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" podUID="f9b85bd6-21c3-457e-81c0-b87c56be9f3e" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.531492 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" event={"ID":"7bead1ee-7a42-45c9-9205-420ae85002f7","Type":"ContainerStarted","Data":"2def0fc8d536163fbe5d95a6ff026b1b5f3001b973cdf38172c1a76a731862e2"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.531855 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.533772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" event={"ID":"2158b3e8-3c08-4e7c-9b28-cd32534135a1","Type":"ContainerStarted","Data":"0e3cb9710ec08c0261c2017ab532853544f6801bb57a620552a78a90a67cec1d"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.534593 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" Nov 26 06:38:55 crc kubenswrapper[4775]: E1126 06:38:55.534869 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:d65dbfc956e9cf376f3c48fc3a0942cb7306b5164f898c40d1efca106df81db7\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" podUID="2158b3e8-3c08-4e7c-9b28-cd32534135a1" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.543291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" event={"ID":"1a1d393e-2d6c-425c-9e6f-068d4c8ee090","Type":"ContainerStarted","Data":"55dfa4c0030fb28cd95486b7407a0f925da0cb7a4ef61a315697a990f0632676"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.544216 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.547620 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.550191 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" event={"ID":"ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7","Type":"ContainerStarted","Data":"34ed23774e568579ecc465a4eb32d9aebf6963ea1419fc424bfd929064431542"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.550509 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.552270 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.552733 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" event={"ID":"153b27e4-d948-4ff6-9d6a-bfbcebd17cad","Type":"ContainerStarted","Data":"c154356e9c14edc373c906db5bce20fa31bd8fb1823eb23218a20a78444dec91"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.553580 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.571537 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.578083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" event={"ID":"7ac95117-0ba0-4c51-afbf-320322b8ebfc","Type":"ContainerStarted","Data":"1826e11d4a4b40137538e166950b990eb1cc9b33e3f9c6081a04d6c219bc6f9a"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.601412 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" event={"ID":"faa5a2e8-ca2b-42e1-ac9a-3580111707ac","Type":"ContainerStarted","Data":"90a8ff4d3308501acb1421c0be5284ccfdecdc32caa369ef22c951c5db45a68f"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.602155 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.603834 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" event={"ID":"cf2135a7-5206-498c-9391-379fae5f4bfa","Type":"ContainerStarted","Data":"98763255bfc0aa6819a30f76f4522ae9d1d15e22efaf58898d1e38bec236fa80"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.603858 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" event={"ID":"cf2135a7-5206-498c-9391-379fae5f4bfa","Type":"ContainerStarted","Data":"7e8043ab0c7d11cec7bd521c252b93f4021a4e188d28d5bfa30478986d19a867"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.604193 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.616019 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6fdcddb789-q97hn" podStartSLOduration=3.469635897 podStartE2EDuration="19.616002812s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.271411294 +0000 UTC m=+801.632715256" lastFinishedPulling="2025-11-26 06:38:54.417778199 +0000 UTC m=+817.779082171" observedRunningTime="2025-11-26 06:38:55.578294682 +0000 UTC m=+818.939598644" watchObservedRunningTime="2025-11-26 06:38:55.616002812 +0000 UTC m=+818.977306764" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.619338 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" event={"ID":"d8ad3c0b-931e-444d-9fcf-23dfbe77bedc","Type":"ContainerStarted","Data":"224c12ce79e1bde46a166753853e0b9b8619113b1b19f87557f608beb161c336"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.621262 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.634086 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.634304 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.653591 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-955677c94-dghfx" podStartSLOduration=2.957465397 podStartE2EDuration="19.653574898s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:37.830844535 +0000 UTC m=+801.192148477" lastFinishedPulling="2025-11-26 06:38:54.526954026 +0000 UTC m=+817.888257978" observedRunningTime="2025-11-26 06:38:55.651290008 +0000 UTC m=+819.012593960" watchObservedRunningTime="2025-11-26 06:38:55.653574898 +0000 UTC m=+819.014878850" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.656044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" event={"ID":"cc351f29-006f-4ee4-80cb-5860f4f824ff","Type":"ContainerStarted","Data":"5c6cb9985e9fcf844a2c7a0934dff704792207507304b001b8eb91230c2cb348"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.656546 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.701335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.723047 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5b77f656f-d7m5d" podStartSLOduration=3.31746496 podStartE2EDuration="19.723025052s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.015478773 +0000 UTC m=+801.376782725" lastFinishedPulling="2025-11-26 06:38:54.421038865 +0000 UTC m=+817.782342817" observedRunningTime="2025-11-26 06:38:55.722074427 +0000 UTC m=+819.083378379" watchObservedRunningTime="2025-11-26 06:38:55.723025052 +0000 UTC m=+819.084329004" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.731985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" event={"ID":"04b644f3-2f69-4cb5-8ade-04f4be267255","Type":"ContainerStarted","Data":"d2e3cb8f1e8f8236f6dee6f280bf04c250f9892bebf61faacd688906d7ac8115"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.732033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" event={"ID":"04b644f3-2f69-4cb5-8ade-04f4be267255","Type":"ContainerStarted","Data":"50f35444346edeed8085661de7318a5213dbc8311cfb510aa2d9d2f545510541"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.732919 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.785403 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-ftsgz" podStartSLOduration=3.74966232 podStartE2EDuration="19.78538561s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.391762094 +0000 UTC m=+801.753066046" lastFinishedPulling="2025-11-26 06:38:54.427485364 +0000 UTC m=+817.788789336" observedRunningTime="2025-11-26 06:38:55.767746916 +0000 UTC m=+819.129050868" watchObservedRunningTime="2025-11-26 06:38:55.78538561 +0000 UTC m=+819.146689562" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.802320 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5d494799bf-jmwr9" podStartSLOduration=3.227254343 podStartE2EDuration="19.802308094s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:37.858885082 +0000 UTC m=+801.220189034" lastFinishedPulling="2025-11-26 06:38:54.433938813 +0000 UTC m=+817.795242785" observedRunningTime="2025-11-26 06:38:55.801055661 +0000 UTC m=+819.162359613" watchObservedRunningTime="2025-11-26 06:38:55.802308094 +0000 UTC m=+819.163612046" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.817107 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" event={"ID":"997d00b8-4cc3-45e9-8af2-2d573eef844a","Type":"ContainerStarted","Data":"2e1b6859565ad8d040046316b19edeebf9019e66d4e61c036eb4904915dd5c3c"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.817924 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.836143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" event={"ID":"021e1511-a060-4d30-af62-872ba84b84c3","Type":"ContainerStarted","Data":"dc7673ec251e4c59e7681f3c495c4359f7390f2543b2aec6ea9b520682c6ef6b"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.836845 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.836934 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" podStartSLOduration=18.836925203 podStartE2EDuration="18.836925203s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:38:55.835959927 +0000 UTC m=+819.197263889" watchObservedRunningTime="2025-11-26 06:38:55.836925203 +0000 UTC m=+819.198229155" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.837335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.846654 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.847503 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" event={"ID":"4f5effa7-3458-4978-8382-0f5b4a17105a","Type":"ContainerStarted","Data":"06f62e6ea06bd0a9ed0934b6217c60ea83f2782e129b9aef4cec1aac647e2475"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.847530 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.858941 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.869349 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6b7f75547b-5whjf" podStartSLOduration=3.618851686 podStartE2EDuration="19.867953698s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.1752873 +0000 UTC m=+801.536591242" lastFinishedPulling="2025-11-26 06:38:54.424389292 +0000 UTC m=+817.785693254" observedRunningTime="2025-11-26 06:38:55.858078638 +0000 UTC m=+819.219382590" watchObservedRunningTime="2025-11-26 06:38:55.867953698 +0000 UTC m=+819.229257650" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.878825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" event={"ID":"d22fe1a8-9183-4ca4-ae47-60d3e77f2339","Type":"ContainerStarted","Data":"753cebf7904b6f6d3c59616acdd0d590585e7657d1907eec8016afc2b38fb23f"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.878859 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" event={"ID":"d22fe1a8-9183-4ca4-ae47-60d3e77f2339","Type":"ContainerStarted","Data":"e2ed1e49409fb5a317a04aab0a429cdc7527bcfaccab1c4f4a1b453f76de389f"} Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.879448 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.909476 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-589cbd6b5b-sdz7q" podStartSLOduration=3.336557281 podStartE2EDuration="19.909461097s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:37.846196508 +0000 UTC m=+801.207500460" lastFinishedPulling="2025-11-26 06:38:54.419100324 +0000 UTC m=+817.780404276" observedRunningTime="2025-11-26 06:38:55.893425986 +0000 UTC m=+819.254729938" watchObservedRunningTime="2025-11-26 06:38:55.909461097 +0000 UTC m=+819.270765049" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.910496 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7b64f4fb85-27cqk" podStartSLOduration=3.540530418 podStartE2EDuration="19.910491425s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.037332917 +0000 UTC m=+801.398636869" lastFinishedPulling="2025-11-26 06:38:54.407293914 +0000 UTC m=+817.768597876" observedRunningTime="2025-11-26 06:38:55.907912397 +0000 UTC m=+819.269216349" watchObservedRunningTime="2025-11-26 06:38:55.910491425 +0000 UTC m=+819.271795377" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.949080 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" podStartSLOduration=3.149987191 podStartE2EDuration="18.949066647s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.585194523 +0000 UTC m=+801.946498465" lastFinishedPulling="2025-11-26 06:38:54.384273939 +0000 UTC m=+817.745577921" observedRunningTime="2025-11-26 06:38:55.933657673 +0000 UTC m=+819.294961635" watchObservedRunningTime="2025-11-26 06:38:55.949066647 +0000 UTC m=+819.310370599" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.950367 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5d499bf58b-5hvj7" podStartSLOduration=3.838626285 podStartE2EDuration="19.950361291s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.272459641 +0000 UTC m=+801.633763593" lastFinishedPulling="2025-11-26 06:38:54.384194617 +0000 UTC m=+817.745498599" observedRunningTime="2025-11-26 06:38:55.947697192 +0000 UTC m=+819.309001144" watchObservedRunningTime="2025-11-26 06:38:55.950361291 +0000 UTC m=+819.311665243" Nov 26 06:38:55 crc kubenswrapper[4775]: I1126 06:38:55.999805 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b4567c7cf-4fxcp" podStartSLOduration=3.418260457 podStartE2EDuration="19.999787129s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:37.844620587 +0000 UTC m=+801.205924539" lastFinishedPulling="2025-11-26 06:38:54.426147259 +0000 UTC m=+817.787451211" observedRunningTime="2025-11-26 06:38:55.969309479 +0000 UTC m=+819.330613431" watchObservedRunningTime="2025-11-26 06:38:55.999787129 +0000 UTC m=+819.361091081" Nov 26 06:38:56 crc kubenswrapper[4775]: I1126 06:38:56.034413 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" podStartSLOduration=3.045280142 podStartE2EDuration="19.034396688s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.395824891 +0000 UTC m=+801.757128843" lastFinishedPulling="2025-11-26 06:38:54.384941397 +0000 UTC m=+817.746245389" observedRunningTime="2025-11-26 06:38:56.034121561 +0000 UTC m=+819.395425503" watchObservedRunningTime="2025-11-26 06:38:56.034396688 +0000 UTC m=+819.395700640" Nov 26 06:38:56 crc kubenswrapper[4775]: E1126 06:38:56.892951 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:225958f250a1075b69439d776a13acc45c78695c21abda23600fb53ca1640423\\\"\"" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" podUID="f9b85bd6-21c3-457e-81c0-b87c56be9f3e" Nov 26 06:38:56 crc kubenswrapper[4775]: E1126 06:38:56.894075 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:d65dbfc956e9cf376f3c48fc3a0942cb7306b5164f898c40d1efca106df81db7\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" podUID="2158b3e8-3c08-4e7c-9b28-cd32534135a1" Nov 26 06:38:57 crc kubenswrapper[4775]: I1126 06:38:57.549278 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.930819 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" event={"ID":"5e867f5f-7847-4469-95cb-605042211f56","Type":"ContainerStarted","Data":"56d899a4ed0536147759f18c580c6bcbc913f221772a227caacb856c71e5a107"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.931377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" event={"ID":"5e867f5f-7847-4469-95cb-605042211f56","Type":"ContainerStarted","Data":"1f1ce35ab69cc2678ed89d4e8c4af69d64b657ae9f9a747239395b44c8b9f7ec"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.932518 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.934146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" event={"ID":"2e76307b-7079-41d7-a66e-35f03cc1a1bd","Type":"ContainerStarted","Data":"fb7d6ae7fc0c53738a45eb69bc218e45e906c7b06513efd08c6e7b5fccc71c9a"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.936258 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" event={"ID":"e72bf30c-6c74-46a6-b215-ac5b1f33c144","Type":"ContainerStarted","Data":"da4a48ee7d317fa7b609395b18c7cf277eaa7784e64cb8ee758e921de29d1df6"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.936293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" event={"ID":"e72bf30c-6c74-46a6-b215-ac5b1f33c144","Type":"ContainerStarted","Data":"a3bd422bb3fc047e027eb17a91f6d51ff0c639b7a7ac8a69c72a5b59fcdeddc8"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.936773 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.938422 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" event={"ID":"85c90cee-6304-4ab9-957c-0bd91411403f","Type":"ContainerStarted","Data":"2a47b8e786d92c0753ebd437e1cc43703525e4239ca8f81d0f4e701438e06a27"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.942242 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" event={"ID":"0fa640fb-a493-4f84-bfcf-a3824ddc061f","Type":"ContainerStarted","Data":"1296c8336b0926ec5d63a14a6b00c960404d25b34085ea59257c26f045afab32"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.942288 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" event={"ID":"0fa640fb-a493-4f84-bfcf-a3824ddc061f","Type":"ContainerStarted","Data":"4df8e0a10cbdbc4b09e35796ba8d52afd3c3cf2de7ecb8c922718ac22e5dff14"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.942899 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.944410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" event={"ID":"7ac95117-0ba0-4c51-afbf-320322b8ebfc","Type":"ContainerStarted","Data":"0ce74b849cad7af040db3af1b615c129a6b7bb79aa6234efe09fa8515f5cd9d3"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.944440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" event={"ID":"7ac95117-0ba0-4c51-afbf-320322b8ebfc","Type":"ContainerStarted","Data":"531b094dec2b328565927e417f1e3122f60187dda60e9be9ba9a8f5ec6ad0d93"} Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.944976 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.956700 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" podStartSLOduration=2.214240661 podStartE2EDuration="24.956680063s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.590930694 +0000 UTC m=+801.952234646" lastFinishedPulling="2025-11-26 06:39:01.333370086 +0000 UTC m=+824.694674048" observedRunningTime="2025-11-26 06:39:01.952812651 +0000 UTC m=+825.314116633" watchObservedRunningTime="2025-11-26 06:39:01.956680063 +0000 UTC m=+825.317984015" Nov 26 06:39:01 crc kubenswrapper[4775]: I1126 06:39:01.987834 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-r7dqc" podStartSLOduration=2.099830236 podStartE2EDuration="24.98781625s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.585282925 +0000 UTC m=+801.946586877" lastFinishedPulling="2025-11-26 06:39:01.473268939 +0000 UTC m=+824.834572891" observedRunningTime="2025-11-26 06:39:01.982697896 +0000 UTC m=+825.344001848" watchObservedRunningTime="2025-11-26 06:39:01.98781625 +0000 UTC m=+825.349120202" Nov 26 06:39:02 crc kubenswrapper[4775]: I1126 06:39:02.024132 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" podStartSLOduration=18.673920563 podStartE2EDuration="25.024111723s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:54.982513998 +0000 UTC m=+818.343817950" lastFinishedPulling="2025-11-26 06:39:01.332705168 +0000 UTC m=+824.694009110" observedRunningTime="2025-11-26 06:39:02.018074265 +0000 UTC m=+825.379378237" watchObservedRunningTime="2025-11-26 06:39:02.024111723 +0000 UTC m=+825.385415675" Nov 26 06:39:02 crc kubenswrapper[4775]: I1126 06:39:02.035930 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" podStartSLOduration=2.161956978 podStartE2EDuration="25.035913603s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.47652513 +0000 UTC m=+801.837829072" lastFinishedPulling="2025-11-26 06:39:01.350481745 +0000 UTC m=+824.711785697" observedRunningTime="2025-11-26 06:39:02.032433182 +0000 UTC m=+825.393737144" watchObservedRunningTime="2025-11-26 06:39:02.035913603 +0000 UTC m=+825.397217555" Nov 26 06:39:02 crc kubenswrapper[4775]: I1126 06:39:02.052339 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" podStartSLOduration=2.280989283 podStartE2EDuration="25.052324594s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.589776213 +0000 UTC m=+801.951080165" lastFinishedPulling="2025-11-26 06:39:01.361111524 +0000 UTC m=+824.722415476" observedRunningTime="2025-11-26 06:39:02.048423272 +0000 UTC m=+825.409727234" watchObservedRunningTime="2025-11-26 06:39:02.052324594 +0000 UTC m=+825.413628546" Nov 26 06:39:02 crc kubenswrapper[4775]: I1126 06:39:02.078822 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-66f4dd4bc7-6jrgg" podStartSLOduration=13.807561257 podStartE2EDuration="26.078800819s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.272939514 +0000 UTC m=+801.634243486" lastFinishedPulling="2025-11-26 06:38:50.544179086 +0000 UTC m=+813.905483048" observedRunningTime="2025-11-26 06:39:02.070001428 +0000 UTC m=+825.431305380" watchObservedRunningTime="2025-11-26 06:39:02.078800819 +0000 UTC m=+825.440104771" Nov 26 06:39:02 crc kubenswrapper[4775]: E1126 06:39:02.450516 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podUID="523ab37c-4bb4-4d00-8134-ad63c2833907" Nov 26 06:39:02 crc kubenswrapper[4775]: I1126 06:39:02.954849 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" event={"ID":"523ab37c-4bb4-4d00-8134-ad63c2833907","Type":"ContainerStarted","Data":"92165fda03262fefb251438c45d562c78426ac47668166d835e9c18b5452d70e"} Nov 26 06:39:02 crc kubenswrapper[4775]: E1126 06:39:02.957848 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:210517b918e30df1c95fc7d961c8e57e9a9d1cc2b9fe7eb4dad2034dd53a90aa\\\"\"" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podUID="523ab37c-4bb4-4d00-8134-ad63c2833907" Nov 26 06:39:03 crc kubenswrapper[4775]: I1126 06:39:03.511141 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-646778448f-7d5mc" Nov 26 06:39:07 crc kubenswrapper[4775]: I1126 06:39:07.589645 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-64cdc6ff96-plpx4" Nov 26 06:39:07 crc kubenswrapper[4775]: I1126 06:39:07.640654 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-56897c768d-ksct6" Nov 26 06:39:07 crc kubenswrapper[4775]: I1126 06:39:07.694551 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d77b94747-2s825" Nov 26 06:39:07 crc kubenswrapper[4775]: I1126 06:39:07.710312 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-fkb27" Nov 26 06:39:07 crc kubenswrapper[4775]: I1126 06:39:07.887060 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-656dcb59d4-65qdq" Nov 26 06:39:08 crc kubenswrapper[4775]: I1126 06:39:08.923085 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:39:08 crc kubenswrapper[4775]: I1126 06:39:08.932736 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8483a9f8-d70d-497d-ae02-c6b426a49306-cert\") pod \"infra-operator-controller-manager-57548d458d-xrzlh\" (UID: \"8483a9f8-d70d-497d-ae02-c6b426a49306\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:39:09 crc kubenswrapper[4775]: I1126 06:39:09.026179 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:39:09 crc kubenswrapper[4775]: I1126 06:39:09.263910 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh"] Nov 26 06:39:09 crc kubenswrapper[4775]: W1126 06:39:09.270327 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8483a9f8_d70d_497d_ae02_c6b426a49306.slice/crio-960b6d75946d95432a6560325288e54ff9961ef8d36f586db6e4276ebdb7d93c WatchSource:0}: Error finding container 960b6d75946d95432a6560325288e54ff9961ef8d36f586db6e4276ebdb7d93c: Status 404 returned error can't find the container with id 960b6d75946d95432a6560325288e54ff9961ef8d36f586db6e4276ebdb7d93c Nov 26 06:39:10 crc kubenswrapper[4775]: I1126 06:39:10.005013 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" event={"ID":"8483a9f8-d70d-497d-ae02-c6b426a49306","Type":"ContainerStarted","Data":"960b6d75946d95432a6560325288e54ff9961ef8d36f586db6e4276ebdb7d93c"} Nov 26 06:39:13 crc kubenswrapper[4775]: I1126 06:39:13.205657 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv" Nov 26 06:39:14 crc kubenswrapper[4775]: I1126 06:39:14.039924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" event={"ID":"f9b85bd6-21c3-457e-81c0-b87c56be9f3e","Type":"ContainerStarted","Data":"e4818bda5bcf6d3cf96fc053e3dd767042b2d0a679befb1ffce853b0528c2c48"} Nov 26 06:39:14 crc kubenswrapper[4775]: I1126 06:39:14.040184 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:39:14 crc kubenswrapper[4775]: I1126 06:39:14.042940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" event={"ID":"2158b3e8-3c08-4e7c-9b28-cd32534135a1","Type":"ContainerStarted","Data":"b58000c59c2f508cb284e6f801bebe439f82eaf53610042e1c288a57b24c7040"} Nov 26 06:39:14 crc kubenswrapper[4775]: I1126 06:39:14.043167 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:39:14 crc kubenswrapper[4775]: I1126 06:39:14.053910 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" podStartSLOduration=2.5524244400000002 podStartE2EDuration="37.053901387s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.468210411 +0000 UTC m=+801.829514363" lastFinishedPulling="2025-11-26 06:39:12.969687338 +0000 UTC m=+836.330991310" observedRunningTime="2025-11-26 06:39:14.052483559 +0000 UTC m=+837.413787521" watchObservedRunningTime="2025-11-26 06:39:14.053901387 +0000 UTC m=+837.415205339" Nov 26 06:39:14 crc kubenswrapper[4775]: I1126 06:39:14.069077 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" podStartSLOduration=3.184672343 podStartE2EDuration="38.069048704s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.011243712 +0000 UTC m=+801.372547664" lastFinishedPulling="2025-11-26 06:39:12.895620073 +0000 UTC m=+836.256924025" observedRunningTime="2025-11-26 06:39:14.066894868 +0000 UTC m=+837.428198850" watchObservedRunningTime="2025-11-26 06:39:14.069048704 +0000 UTC m=+837.430352696" Nov 26 06:39:15 crc kubenswrapper[4775]: I1126 06:39:15.068273 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" event={"ID":"8483a9f8-d70d-497d-ae02-c6b426a49306","Type":"ContainerStarted","Data":"bcbd4b8e5e49dd109b22bc123838cbefacf4ec7e4a8a171a4bf888641304cbcd"} Nov 26 06:39:15 crc kubenswrapper[4775]: I1126 06:39:15.068659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" event={"ID":"8483a9f8-d70d-497d-ae02-c6b426a49306","Type":"ContainerStarted","Data":"2ce5f183ae9f4fabd787a009603cfe9c673b80f862e09ccce58ecb5fd443d565"} Nov 26 06:39:15 crc kubenswrapper[4775]: I1126 06:39:15.068767 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:39:15 crc kubenswrapper[4775]: I1126 06:39:15.107223 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" podStartSLOduration=34.272274579 podStartE2EDuration="39.107202923s" podCreationTimestamp="2025-11-26 06:38:36 +0000 UTC" firstStartedPulling="2025-11-26 06:39:09.272110978 +0000 UTC m=+832.633414930" lastFinishedPulling="2025-11-26 06:39:14.107039312 +0000 UTC m=+837.468343274" observedRunningTime="2025-11-26 06:39:15.097892379 +0000 UTC m=+838.459196371" watchObservedRunningTime="2025-11-26 06:39:15.107202923 +0000 UTC m=+838.468506875" Nov 26 06:39:19 crc kubenswrapper[4775]: I1126 06:39:19.032046 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-xrzlh" Nov 26 06:39:21 crc kubenswrapper[4775]: I1126 06:39:21.130161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" event={"ID":"523ab37c-4bb4-4d00-8134-ad63c2833907","Type":"ContainerStarted","Data":"93ef7823fedf565ce0df3708026d905c63984a93d3658814d92bc93112c382cc"} Nov 26 06:39:21 crc kubenswrapper[4775]: I1126 06:39:21.131221 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:39:21 crc kubenswrapper[4775]: I1126 06:39:21.173575 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" podStartSLOduration=2.530327407 podStartE2EDuration="44.173550208s" podCreationTimestamp="2025-11-26 06:38:37 +0000 UTC" firstStartedPulling="2025-11-26 06:38:38.590940584 +0000 UTC m=+801.952244526" lastFinishedPulling="2025-11-26 06:39:20.234163375 +0000 UTC m=+843.595467327" observedRunningTime="2025-11-26 06:39:21.173304992 +0000 UTC m=+844.534609014" watchObservedRunningTime="2025-11-26 06:39:21.173550208 +0000 UTC m=+844.534854190" Nov 26 06:39:27 crc kubenswrapper[4775]: I1126 06:39:27.345480 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67cb4dc6d4-gdk4v" Nov 26 06:39:27 crc kubenswrapper[4775]: I1126 06:39:27.623450 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-57988cc5b5-ffw7r" Nov 26 06:39:27 crc kubenswrapper[4775]: I1126 06:39:27.803588 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd6c7f4c8-8sjts" Nov 26 06:39:48 crc kubenswrapper[4775]: I1126 06:39:48.994101 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kbgn7"] Nov 26 06:39:48 crc kubenswrapper[4775]: I1126 06:39:48.996042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:48 crc kubenswrapper[4775]: I1126 06:39:48.998583 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 26 06:39:48 crc kubenswrapper[4775]: I1126 06:39:48.998697 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 26 06:39:48 crc kubenswrapper[4775]: I1126 06:39:48.998853 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rpknd" Nov 26 06:39:48 crc kubenswrapper[4775]: I1126 06:39:48.999106 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.011294 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kbgn7"] Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.053304 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h8ljd"] Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.054494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.059987 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.064628 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h8ljd"] Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.084247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.084301 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htjqw\" (UniqueName: \"kubernetes.io/projected/d6a7ce1d-2609-4d78-9565-c900247f40d6-kube-api-access-htjqw\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.084397 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-config\") pod \"dnsmasq-dns-675f4bcbfc-kbgn7\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.084451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf5bg\" (UniqueName: \"kubernetes.io/projected/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-kube-api-access-pf5bg\") pod \"dnsmasq-dns-675f4bcbfc-kbgn7\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.084484 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-config\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.100451 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-258zm"] Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.104836 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.113009 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-258zm"] Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.185951 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-config\") pod \"dnsmasq-dns-675f4bcbfc-kbgn7\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-utilities\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186060 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf5bg\" (UniqueName: \"kubernetes.io/projected/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-kube-api-access-pf5bg\") pod \"dnsmasq-dns-675f4bcbfc-kbgn7\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-config\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186135 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htjqw\" (UniqueName: \"kubernetes.io/projected/d6a7ce1d-2609-4d78-9565-c900247f40d6-kube-api-access-htjqw\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-catalog-content\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.186178 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k2sl\" (UniqueName: \"kubernetes.io/projected/193ae057-ca33-4654-95f8-5453f0938ce7-kube-api-access-2k2sl\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.187079 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.187206 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-config\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.188254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-config\") pod \"dnsmasq-dns-675f4bcbfc-kbgn7\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.205510 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htjqw\" (UniqueName: \"kubernetes.io/projected/d6a7ce1d-2609-4d78-9565-c900247f40d6-kube-api-access-htjqw\") pod \"dnsmasq-dns-78dd6ddcc-h8ljd\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.207178 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf5bg\" (UniqueName: \"kubernetes.io/projected/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-kube-api-access-pf5bg\") pod \"dnsmasq-dns-675f4bcbfc-kbgn7\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.287361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-utilities\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.287447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-catalog-content\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.287480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k2sl\" (UniqueName: \"kubernetes.io/projected/193ae057-ca33-4654-95f8-5453f0938ce7-kube-api-access-2k2sl\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.288426 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-utilities\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.288660 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-catalog-content\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.308508 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k2sl\" (UniqueName: \"kubernetes.io/projected/193ae057-ca33-4654-95f8-5453f0938ce7-kube-api-access-2k2sl\") pod \"redhat-operators-258zm\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.316294 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.373666 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.420797 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.751080 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kbgn7"] Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.759788 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.852478 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h8ljd"] Nov 26 06:39:49 crc kubenswrapper[4775]: W1126 06:39:49.862093 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6a7ce1d_2609_4d78_9565_c900247f40d6.slice/crio-045f6b9816c19347cc664640593a5ba60a0d064f59cbfddabca48372b7f875b9 WatchSource:0}: Error finding container 045f6b9816c19347cc664640593a5ba60a0d064f59cbfddabca48372b7f875b9: Status 404 returned error can't find the container with id 045f6b9816c19347cc664640593a5ba60a0d064f59cbfddabca48372b7f875b9 Nov 26 06:39:49 crc kubenswrapper[4775]: I1126 06:39:49.939495 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-258zm"] Nov 26 06:39:49 crc kubenswrapper[4775]: W1126 06:39:49.939784 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod193ae057_ca33_4654_95f8_5453f0938ce7.slice/crio-2d9685ad860baca1304daaedf298ad35fbdde32e9a306336b7ff2162db30c5d3 WatchSource:0}: Error finding container 2d9685ad860baca1304daaedf298ad35fbdde32e9a306336b7ff2162db30c5d3: Status 404 returned error can't find the container with id 2d9685ad860baca1304daaedf298ad35fbdde32e9a306336b7ff2162db30c5d3 Nov 26 06:39:50 crc kubenswrapper[4775]: I1126 06:39:50.389345 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" event={"ID":"d6a7ce1d-2609-4d78-9565-c900247f40d6","Type":"ContainerStarted","Data":"045f6b9816c19347cc664640593a5ba60a0d064f59cbfddabca48372b7f875b9"} Nov 26 06:39:50 crc kubenswrapper[4775]: I1126 06:39:50.392352 4775 generic.go:334] "Generic (PLEG): container finished" podID="193ae057-ca33-4654-95f8-5453f0938ce7" containerID="446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945" exitCode=0 Nov 26 06:39:50 crc kubenswrapper[4775]: I1126 06:39:50.392406 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-258zm" event={"ID":"193ae057-ca33-4654-95f8-5453f0938ce7","Type":"ContainerDied","Data":"446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945"} Nov 26 06:39:50 crc kubenswrapper[4775]: I1126 06:39:50.392424 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-258zm" event={"ID":"193ae057-ca33-4654-95f8-5453f0938ce7","Type":"ContainerStarted","Data":"2d9685ad860baca1304daaedf298ad35fbdde32e9a306336b7ff2162db30c5d3"} Nov 26 06:39:50 crc kubenswrapper[4775]: I1126 06:39:50.393488 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" event={"ID":"f6381ad9-d93b-4f52-a5ae-4d694f4db96d","Type":"ContainerStarted","Data":"c9b13f23182a476d3f0c1105cc145cb53898f751c20668a5a117c09f591fea09"} Nov 26 06:39:51 crc kubenswrapper[4775]: I1126 06:39:51.420313 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:39:51 crc kubenswrapper[4775]: I1126 06:39:51.420351 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.155928 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kbgn7"] Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.184021 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tfrkk"] Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.185451 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.196778 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tfrkk"] Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.233637 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-config\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.233759 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrngt\" (UniqueName: \"kubernetes.io/projected/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-kube-api-access-xrngt\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.233788 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.335198 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-config\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.335278 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrngt\" (UniqueName: \"kubernetes.io/projected/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-kube-api-access-xrngt\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.335296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.336005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.336500 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-config\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.371849 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrngt\" (UniqueName: \"kubernetes.io/projected/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-kube-api-access-xrngt\") pod \"dnsmasq-dns-666b6646f7-tfrkk\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.430528 4775 generic.go:334] "Generic (PLEG): container finished" podID="193ae057-ca33-4654-95f8-5453f0938ce7" containerID="fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1" exitCode=0 Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.430585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-258zm" event={"ID":"193ae057-ca33-4654-95f8-5453f0938ce7","Type":"ContainerDied","Data":"fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1"} Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.438961 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h8ljd"] Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.467241 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nfsh5"] Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.468441 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.480475 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nfsh5"] Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.516530 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.537735 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br7q6\" (UniqueName: \"kubernetes.io/projected/01958f53-a2a9-4783-9669-abd002d57a66-kube-api-access-br7q6\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.537816 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.537842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-config\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.639688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br7q6\" (UniqueName: \"kubernetes.io/projected/01958f53-a2a9-4783-9669-abd002d57a66-kube-api-access-br7q6\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.639783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.639804 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-config\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.640559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-config\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.640956 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.666624 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br7q6\" (UniqueName: \"kubernetes.io/projected/01958f53-a2a9-4783-9669-abd002d57a66-kube-api-access-br7q6\") pod \"dnsmasq-dns-57d769cc4f-nfsh5\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:52 crc kubenswrapper[4775]: I1126 06:39:52.792380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.325096 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.326380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.330176 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.330769 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xrgm4" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.331745 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.334193 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.334432 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.335213 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.335523 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351249 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351372 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351407 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351460 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-config-data\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351499 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63f863dc-36a4-4325-b520-e26b3dca309e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351522 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351657 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr5v9\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-kube-api-access-jr5v9\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351746 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.351834 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63f863dc-36a4-4325-b520-e26b3dca309e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.354960 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.453015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63f863dc-36a4-4325-b520-e26b3dca309e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.453861 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454051 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr5v9\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-kube-api-access-jr5v9\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63f863dc-36a4-4325-b520-e26b3dca309e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454288 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454295 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454313 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454348 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454387 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-config-data\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.454754 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.457058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.457139 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.459868 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-config-data\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.462263 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.463249 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.463531 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63f863dc-36a4-4325-b520-e26b3dca309e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.464363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.465654 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63f863dc-36a4-4325-b520-e26b3dca309e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.470984 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr5v9\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-kube-api-access-jr5v9\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.479863 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.582167 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.583280 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.585216 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.585263 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.585302 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.585569 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.585594 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.585680 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qc8lq" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.589081 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.593460 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.658407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.758977 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-979q2\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-kube-api-access-979q2\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759052 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759085 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92a5651e-7175-45ca-a69b-dc19a11f2943-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759285 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759348 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759376 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759484 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92a5651e-7175-45ca-a69b-dc19a11f2943-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759547 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759605 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.759631 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.860890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-979q2\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-kube-api-access-979q2\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.860951 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.860972 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.860990 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92a5651e-7175-45ca-a69b-dc19a11f2943-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861067 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92a5651e-7175-45ca-a69b-dc19a11f2943-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861302 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861415 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861446 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.861463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.863057 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.863618 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.864346 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.864876 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.865752 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.867070 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92a5651e-7175-45ca-a69b-dc19a11f2943-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.867415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.869191 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.882244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92a5651e-7175-45ca-a69b-dc19a11f2943-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.887642 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.895381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-979q2\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-kube-api-access-979q2\") pod \"rabbitmq-cell1-server-0\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:53 crc kubenswrapper[4775]: I1126 06:39:53.918280 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.025542 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.076017 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.079563 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.097409 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-8frrj" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.097910 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.098135 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.116357 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.117840 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188483 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0028de9f-a7db-47c0-a552-e9f518435458-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2wzh\" (UniqueName: \"kubernetes.io/projected/0028de9f-a7db-47c0-a552-e9f518435458-kube-api-access-s2wzh\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188588 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0028de9f-a7db-47c0-a552-e9f518435458-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188686 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-kolla-config\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.188849 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-config-data-default\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.189066 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0028de9f-a7db-47c0-a552-e9f518435458-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.290808 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0028de9f-a7db-47c0-a552-e9f518435458-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.290859 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2wzh\" (UniqueName: \"kubernetes.io/projected/0028de9f-a7db-47c0-a552-e9f518435458-kube-api-access-s2wzh\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.290892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0028de9f-a7db-47c0-a552-e9f518435458-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.290941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.290967 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.290994 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-kolla-config\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.291047 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-config-data-default\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.291094 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0028de9f-a7db-47c0-a552-e9f518435458-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.291457 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.291507 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0028de9f-a7db-47c0-a552-e9f518435458-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.292182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-kolla-config\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.292557 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-config-data-default\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.292793 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0028de9f-a7db-47c0-a552-e9f518435458-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.297938 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0028de9f-a7db-47c0-a552-e9f518435458-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.302922 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0028de9f-a7db-47c0-a552-e9f518435458-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.319281 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.328352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2wzh\" (UniqueName: \"kubernetes.io/projected/0028de9f-a7db-47c0-a552-e9f518435458-kube-api-access-s2wzh\") pod \"openstack-galera-0\" (UID: \"0028de9f-a7db-47c0-a552-e9f518435458\") " pod="openstack/openstack-galera-0" Nov 26 06:39:55 crc kubenswrapper[4775]: I1126 06:39:55.426910 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.403815 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.407849 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.410978 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-b8dht" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.415235 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.418998 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.419311 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.419517 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.514814 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.514863 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpt4t\" (UniqueName: \"kubernetes.io/projected/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-kube-api-access-hpt4t\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.514905 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.515047 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.515152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.515182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.515232 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.515261 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616341 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616370 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616395 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616475 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpt4t\" (UniqueName: \"kubernetes.io/projected/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-kube-api-access-hpt4t\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616502 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616548 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.616683 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.617474 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.617476 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.618404 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.618965 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.622325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.622699 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.634905 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpt4t\" (UniqueName: \"kubernetes.io/projected/7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c-kube-api-access-hpt4t\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.671361 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c\") " pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.741336 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.842287 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.843476 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.845514 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.845659 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.845743 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hkn84" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.854595 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.923393 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bf022b43-4760-4e21-8f5b-2558a69ede14-kolla-config\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.923459 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf022b43-4760-4e21-8f5b-2558a69ede14-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.923500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf022b43-4760-4e21-8f5b-2558a69ede14-config-data\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.923552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl9tc\" (UniqueName: \"kubernetes.io/projected/bf022b43-4760-4e21-8f5b-2558a69ede14-kube-api-access-sl9tc\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:56 crc kubenswrapper[4775]: I1126 06:39:56.923575 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf022b43-4760-4e21-8f5b-2558a69ede14-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.024431 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bf022b43-4760-4e21-8f5b-2558a69ede14-kolla-config\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.024493 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf022b43-4760-4e21-8f5b-2558a69ede14-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.024544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf022b43-4760-4e21-8f5b-2558a69ede14-config-data\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.024596 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl9tc\" (UniqueName: \"kubernetes.io/projected/bf022b43-4760-4e21-8f5b-2558a69ede14-kube-api-access-sl9tc\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.024617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf022b43-4760-4e21-8f5b-2558a69ede14-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.025201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bf022b43-4760-4e21-8f5b-2558a69ede14-kolla-config\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.025761 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf022b43-4760-4e21-8f5b-2558a69ede14-config-data\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.030353 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf022b43-4760-4e21-8f5b-2558a69ede14-memcached-tls-certs\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.030361 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf022b43-4760-4e21-8f5b-2558a69ede14-combined-ca-bundle\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.073174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl9tc\" (UniqueName: \"kubernetes.io/projected/bf022b43-4760-4e21-8f5b-2558a69ede14-kube-api-access-sl9tc\") pod \"memcached-0\" (UID: \"bf022b43-4760-4e21-8f5b-2558a69ede14\") " pod="openstack/memcached-0" Nov 26 06:39:57 crc kubenswrapper[4775]: I1126 06:39:57.173157 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.585712 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.586633 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.589394 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-xj5gt" Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.595029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.650688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gnrp\" (UniqueName: \"kubernetes.io/projected/633ccd9d-4b8e-4797-89f9-7412c416b35b-kube-api-access-4gnrp\") pod \"kube-state-metrics-0\" (UID: \"633ccd9d-4b8e-4797-89f9-7412c416b35b\") " pod="openstack/kube-state-metrics-0" Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.752636 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gnrp\" (UniqueName: \"kubernetes.io/projected/633ccd9d-4b8e-4797-89f9-7412c416b35b-kube-api-access-4gnrp\") pod \"kube-state-metrics-0\" (UID: \"633ccd9d-4b8e-4797-89f9-7412c416b35b\") " pod="openstack/kube-state-metrics-0" Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.772788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gnrp\" (UniqueName: \"kubernetes.io/projected/633ccd9d-4b8e-4797-89f9-7412c416b35b-kube-api-access-4gnrp\") pod \"kube-state-metrics-0\" (UID: \"633ccd9d-4b8e-4797-89f9-7412c416b35b\") " pod="openstack/kube-state-metrics-0" Nov 26 06:39:58 crc kubenswrapper[4775]: I1126 06:39:58.965076 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 06:39:59 crc kubenswrapper[4775]: I1126 06:39:59.147442 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tfrkk"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.066787 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hb7d2"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.069814 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.080804 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hb7d2"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.109447 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f7rn\" (UniqueName: \"kubernetes.io/projected/46ce0a21-d5df-44fb-96cc-8e008b5e989a-kube-api-access-5f7rn\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.109498 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-catalog-content\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.109584 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-utilities\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.211383 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-utilities\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.211483 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f7rn\" (UniqueName: \"kubernetes.io/projected/46ce0a21-d5df-44fb-96cc-8e008b5e989a-kube-api-access-5f7rn\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.211513 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-catalog-content\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.211859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-utilities\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.212021 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-catalog-content\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.236462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f7rn\" (UniqueName: \"kubernetes.io/projected/46ce0a21-d5df-44fb-96cc-8e008b5e989a-kube-api-access-5f7rn\") pod \"certified-operators-hb7d2\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.299301 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k4mqj"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.300247 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.303188 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.303456 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.303517 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-cpbdl" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.306315 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-4qmsw"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.307809 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.312904 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-run\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.312932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d30382f9-28bd-4ab7-8fec-1f8932b3f804-scripts\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.312954 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-lib\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.312974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-log\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.312996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30382f9-28bd-4ab7-8fec-1f8932b3f804-combined-ca-bundle\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-log-ovn\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313029 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhjp6\" (UniqueName: \"kubernetes.io/projected/d30382f9-28bd-4ab7-8fec-1f8932b3f804-kube-api-access-qhjp6\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313047 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-etc-ovs\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313070 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-run\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313092 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d30382f9-28bd-4ab7-8fec-1f8932b3f804-ovn-controller-tls-certs\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-run-ovn\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313134 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r94vt\" (UniqueName: \"kubernetes.io/projected/64dca619-067a-4f6d-b391-c736d9771a1a-kube-api-access-r94vt\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.313155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64dca619-067a-4f6d-b391-c736d9771a1a-scripts\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.378402 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4mqj"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.388427 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4qmsw"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.403174 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.415070 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-lib\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.415950 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-lib\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416005 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-log\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416030 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30382f9-28bd-4ab7-8fec-1f8932b3f804-combined-ca-bundle\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416047 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-log-ovn\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhjp6\" (UniqueName: \"kubernetes.io/projected/d30382f9-28bd-4ab7-8fec-1f8932b3f804-kube-api-access-qhjp6\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416090 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-etc-ovs\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416125 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-run\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416165 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d30382f9-28bd-4ab7-8fec-1f8932b3f804-ovn-controller-tls-certs\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416187 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-run-ovn\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416219 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r94vt\" (UniqueName: \"kubernetes.io/projected/64dca619-067a-4f6d-b391-c736d9771a1a-kube-api-access-r94vt\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416238 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64dca619-067a-4f6d-b391-c736d9771a1a-scripts\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416282 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-run\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.416298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d30382f9-28bd-4ab7-8fec-1f8932b3f804-scripts\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.417270 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-run\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.418138 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-log\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.418161 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-run-ovn\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.418246 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-var-run\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.418507 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/64dca619-067a-4f6d-b391-c736d9771a1a-etc-ovs\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.418921 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d30382f9-28bd-4ab7-8fec-1f8932b3f804-var-log-ovn\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.419854 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64dca619-067a-4f6d-b391-c736d9771a1a-scripts\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.422617 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30382f9-28bd-4ab7-8fec-1f8932b3f804-combined-ca-bundle\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.423987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d30382f9-28bd-4ab7-8fec-1f8932b3f804-scripts\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.427363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/d30382f9-28bd-4ab7-8fec-1f8932b3f804-ovn-controller-tls-certs\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.433552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r94vt\" (UniqueName: \"kubernetes.io/projected/64dca619-067a-4f6d-b391-c736d9771a1a-kube-api-access-r94vt\") pod \"ovn-controller-ovs-4qmsw\" (UID: \"64dca619-067a-4f6d-b391-c736d9771a1a\") " pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.445704 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhjp6\" (UniqueName: \"kubernetes.io/projected/d30382f9-28bd-4ab7-8fec-1f8932b3f804-kube-api-access-qhjp6\") pod \"ovn-controller-k4mqj\" (UID: \"d30382f9-28bd-4ab7-8fec-1f8932b3f804\") " pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.696257 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.711249 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.721524 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.722739 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.727045 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.727218 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.727459 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-qfwtv" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.727549 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.728344 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.750920 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.922846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.922912 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e61da90a-7acf-4b95-934b-69f09efb5cbc-config\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.922941 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.922965 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vxgf\" (UniqueName: \"kubernetes.io/projected/e61da90a-7acf-4b95-934b-69f09efb5cbc-kube-api-access-7vxgf\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.923014 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e61da90a-7acf-4b95-934b-69f09efb5cbc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.923043 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.923068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e61da90a-7acf-4b95-934b-69f09efb5cbc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:02 crc kubenswrapper[4775]: I1126 06:40:02.923126 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.024763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.024848 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.024892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e61da90a-7acf-4b95-934b-69f09efb5cbc-config\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.024910 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.024927 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vxgf\" (UniqueName: \"kubernetes.io/projected/e61da90a-7acf-4b95-934b-69f09efb5cbc-kube-api-access-7vxgf\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.024977 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e61da90a-7acf-4b95-934b-69f09efb5cbc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.025000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.025032 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e61da90a-7acf-4b95-934b-69f09efb5cbc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.026114 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e61da90a-7acf-4b95-934b-69f09efb5cbc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.027115 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.027214 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e61da90a-7acf-4b95-934b-69f09efb5cbc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.027553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e61da90a-7acf-4b95-934b-69f09efb5cbc-config\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.029828 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.030934 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.043046 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vxgf\" (UniqueName: \"kubernetes.io/projected/e61da90a-7acf-4b95-934b-69f09efb5cbc-kube-api-access-7vxgf\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.060083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e61da90a-7acf-4b95-934b-69f09efb5cbc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.064648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"e61da90a-7acf-4b95-934b-69f09efb5cbc\") " pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:03 crc kubenswrapper[4775]: I1126 06:40:03.337145 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:05 crc kubenswrapper[4775]: W1126 06:40:05.182982 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podceab34dd_6af2_44c4_8e4d_9ad3a92b3c06.slice/crio-e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580 WatchSource:0}: Error finding container e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580: Status 404 returned error can't find the container with id e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580 Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.451514 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9rmnn"] Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.465771 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9rmnn"] Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.465863 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.530654 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" event={"ID":"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06","Type":"ContainerStarted","Data":"e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580"} Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.581546 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g96km\" (UniqueName: \"kubernetes.io/projected/1c87f5c8-5d09-4e93-ad84-a2945745039b-kube-api-access-g96km\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.581777 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-utilities\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.581989 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-catalog-content\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.596406 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.683930 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g96km\" (UniqueName: \"kubernetes.io/projected/1c87f5c8-5d09-4e93-ad84-a2945745039b-kube-api-access-g96km\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.684044 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-utilities\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.684121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-catalog-content\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.684518 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-catalog-content\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.684559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-utilities\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.721646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g96km\" (UniqueName: \"kubernetes.io/projected/1c87f5c8-5d09-4e93-ad84-a2945745039b-kube-api-access-g96km\") pod \"community-operators-9rmnn\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:05 crc kubenswrapper[4775]: I1126 06:40:05.791483 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.125335 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.126560 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.128453 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-mhh85" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.128495 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.129153 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.129505 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.142687 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.292787 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc88dc0e-14a1-479e-9682-bb5c611259a8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.292833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc88dc0e-14a1-479e-9682-bb5c611259a8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.292914 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc88dc0e-14a1-479e-9682-bb5c611259a8-config\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.292953 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2fvq\" (UniqueName: \"kubernetes.io/projected/dc88dc0e-14a1-479e-9682-bb5c611259a8-kube-api-access-p2fvq\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.292978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.293045 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.293068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.293106 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.394922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.394972 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc88dc0e-14a1-479e-9682-bb5c611259a8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc88dc0e-14a1-479e-9682-bb5c611259a8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc88dc0e-14a1-479e-9682-bb5c611259a8-config\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2fvq\" (UniqueName: \"kubernetes.io/projected/dc88dc0e-14a1-479e-9682-bb5c611259a8-kube-api-access-p2fvq\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395207 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.395582 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.396355 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dc88dc0e-14a1-479e-9682-bb5c611259a8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.396864 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc88dc0e-14a1-479e-9682-bb5c611259a8-config\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.397035 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc88dc0e-14a1-479e-9682-bb5c611259a8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.399871 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.400216 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.400809 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc88dc0e-14a1-479e-9682-bb5c611259a8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.411220 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2fvq\" (UniqueName: \"kubernetes.io/projected/dc88dc0e-14a1-479e-9682-bb5c611259a8-kube-api-access-p2fvq\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.417997 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dc88dc0e-14a1-479e-9682-bb5c611259a8\") " pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.445439 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:06 crc kubenswrapper[4775]: W1126 06:40:06.446466 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92a5651e_7175_45ca_a69b_dc19a11f2943.slice/crio-501a9d556dc52a76e2cc46f1734a0285e42c7fcf5e57d92b73892b6703aa0a95 WatchSource:0}: Error finding container 501a9d556dc52a76e2cc46f1734a0285e42c7fcf5e57d92b73892b6703aa0a95: Status 404 returned error can't find the container with id 501a9d556dc52a76e2cc46f1734a0285e42c7fcf5e57d92b73892b6703aa0a95 Nov 26 06:40:06 crc kubenswrapper[4775]: E1126 06:40:06.480711 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 26 06:40:06 crc kubenswrapper[4775]: E1126 06:40:06.480896 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-htjqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-h8ljd_openstack(d6a7ce1d-2609-4d78-9565-c900247f40d6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:40:06 crc kubenswrapper[4775]: E1126 06:40:06.482114 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" podUID="d6a7ce1d-2609-4d78-9565-c900247f40d6" Nov 26 06:40:06 crc kubenswrapper[4775]: E1126 06:40:06.525069 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 26 06:40:06 crc kubenswrapper[4775]: E1126 06:40:06.525205 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pf5bg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-kbgn7_openstack(f6381ad9-d93b-4f52-a5ae-4d694f4db96d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:40:06 crc kubenswrapper[4775]: E1126 06:40:06.526497 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" podUID="f6381ad9-d93b-4f52-a5ae-4d694f4db96d" Nov 26 06:40:06 crc kubenswrapper[4775]: I1126 06:40:06.551490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"92a5651e-7175-45ca-a69b-dc19a11f2943","Type":"ContainerStarted","Data":"501a9d556dc52a76e2cc46f1734a0285e42c7fcf5e57d92b73892b6703aa0a95"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.074439 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nfsh5"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.092502 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 26 06:40:07 crc kubenswrapper[4775]: W1126 06:40:07.100690 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0028de9f_a7db_47c0_a552_e9f518435458.slice/crio-eea167271fdbf41dd29c22a668af9984e6a6e3c5f4ad41b2959c6557ad73449f WatchSource:0}: Error finding container eea167271fdbf41dd29c22a668af9984e6a6e3c5f4ad41b2959c6557ad73449f: Status 404 returned error can't find the container with id eea167271fdbf41dd29c22a668af9984e6a6e3c5f4ad41b2959c6557ad73449f Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.109499 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.140018 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.208507 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pf5bg\" (UniqueName: \"kubernetes.io/projected/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-kube-api-access-pf5bg\") pod \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.208625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-config\") pod \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\" (UID: \"f6381ad9-d93b-4f52-a5ae-4d694f4db96d\") " Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.209600 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-config" (OuterVolumeSpecName: "config") pod "f6381ad9-d93b-4f52-a5ae-4d694f4db96d" (UID: "f6381ad9-d93b-4f52-a5ae-4d694f4db96d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.212514 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-kube-api-access-pf5bg" (OuterVolumeSpecName: "kube-api-access-pf5bg") pod "f6381ad9-d93b-4f52-a5ae-4d694f4db96d" (UID: "f6381ad9-d93b-4f52-a5ae-4d694f4db96d"). InnerVolumeSpecName "kube-api-access-pf5bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.309685 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-dns-svc\") pod \"d6a7ce1d-2609-4d78-9565-c900247f40d6\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.309805 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htjqw\" (UniqueName: \"kubernetes.io/projected/d6a7ce1d-2609-4d78-9565-c900247f40d6-kube-api-access-htjqw\") pod \"d6a7ce1d-2609-4d78-9565-c900247f40d6\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.309908 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-config\") pod \"d6a7ce1d-2609-4d78-9565-c900247f40d6\" (UID: \"d6a7ce1d-2609-4d78-9565-c900247f40d6\") " Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.310308 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pf5bg\" (UniqueName: \"kubernetes.io/projected/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-kube-api-access-pf5bg\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.310324 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6381ad9-d93b-4f52-a5ae-4d694f4db96d-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.310798 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-config" (OuterVolumeSpecName: "config") pod "d6a7ce1d-2609-4d78-9565-c900247f40d6" (UID: "d6a7ce1d-2609-4d78-9565-c900247f40d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.311118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6a7ce1d-2609-4d78-9565-c900247f40d6" (UID: "d6a7ce1d-2609-4d78-9565-c900247f40d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.315191 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6a7ce1d-2609-4d78-9565-c900247f40d6-kube-api-access-htjqw" (OuterVolumeSpecName: "kube-api-access-htjqw") pod "d6a7ce1d-2609-4d78-9565-c900247f40d6" (UID: "d6a7ce1d-2609-4d78-9565-c900247f40d6"). InnerVolumeSpecName "kube-api-access-htjqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.412689 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.413027 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htjqw\" (UniqueName: \"kubernetes.io/projected/d6a7ce1d-2609-4d78-9565-c900247f40d6-kube-api-access-htjqw\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.413041 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6a7ce1d-2609-4d78-9565-c900247f40d6-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.493693 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hb7d2"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.509538 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.524752 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.547823 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.555782 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9rmnn"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.571432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c","Type":"ContainerStarted","Data":"b03d05b835f8acdb1971264877d1c59e1c355b23efd42d1cb07771d9a47b99b6"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.586911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" event={"ID":"f6381ad9-d93b-4f52-a5ae-4d694f4db96d","Type":"ContainerDied","Data":"c9b13f23182a476d3f0c1105cc145cb53898f751c20668a5a117c09f591fea09"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.586996 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kbgn7" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.589704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" event={"ID":"d6a7ce1d-2609-4d78-9565-c900247f40d6","Type":"ContainerDied","Data":"045f6b9816c19347cc664640593a5ba60a0d064f59cbfddabca48372b7f875b9"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.589792 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-h8ljd" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.594158 4775 generic.go:334] "Generic (PLEG): container finished" podID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerID="6ca8a1a2707962f20c8b0b03e1c98798df5ec08fcabad6ff58f7fc18121991c7" exitCode=0 Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.594561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" event={"ID":"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06","Type":"ContainerDied","Data":"6ca8a1a2707962f20c8b0b03e1c98798df5ec08fcabad6ff58f7fc18121991c7"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.606283 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63f863dc-36a4-4325-b520-e26b3dca309e","Type":"ContainerStarted","Data":"f5332c832f6e8c06b03d96af0eb0b1a552b4cd1045106f0c2f4cdb19e13ff2db"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.609075 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-258zm" event={"ID":"193ae057-ca33-4654-95f8-5453f0938ce7","Type":"ContainerStarted","Data":"f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.611128 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0028de9f-a7db-47c0-a552-e9f518435458","Type":"ContainerStarted","Data":"eea167271fdbf41dd29c22a668af9984e6a6e3c5f4ad41b2959c6557ad73449f"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.612389 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerStarted","Data":"c09419cada9c02b8ea8829f97df92f1969477bc59ddc4a4787838ab1e2bbd722"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.614375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rmnn" event={"ID":"1c87f5c8-5d09-4e93-ad84-a2945745039b","Type":"ContainerStarted","Data":"a802c97b1c269018bdf1b9287cfebec903a62a18f5fcd854dfc1a424230657ed"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.627019 4775 generic.go:334] "Generic (PLEG): container finished" podID="01958f53-a2a9-4783-9669-abd002d57a66" containerID="39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85" exitCode=0 Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.627076 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" event={"ID":"01958f53-a2a9-4783-9669-abd002d57a66","Type":"ContainerDied","Data":"39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.627914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" event={"ID":"01958f53-a2a9-4783-9669-abd002d57a66","Type":"ContainerStarted","Data":"af8796a852a66cc2dd686cabb4b075976b7dd9162567689ebd90561e658424dc"} Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.640247 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-258zm" podStartSLOduration=2.57841992 podStartE2EDuration="18.640228063s" podCreationTimestamp="2025-11-26 06:39:49 +0000 UTC" firstStartedPulling="2025-11-26 06:39:50.393780331 +0000 UTC m=+873.755084283" lastFinishedPulling="2025-11-26 06:40:06.455588474 +0000 UTC m=+889.816892426" observedRunningTime="2025-11-26 06:40:07.636858214 +0000 UTC m=+890.998162166" watchObservedRunningTime="2025-11-26 06:40:07.640228063 +0000 UTC m=+891.001532015" Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.850908 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.877104 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4mqj"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.884984 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kbgn7"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.901760 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kbgn7"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.921784 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h8ljd"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.926833 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-h8ljd"] Nov 26 06:40:07 crc kubenswrapper[4775]: I1126 06:40:07.958737 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 26 06:40:08 crc kubenswrapper[4775]: E1126 06:40:08.033566 4775 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 26 06:40:08 crc kubenswrapper[4775]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 26 06:40:08 crc kubenswrapper[4775]: > podSandboxID="e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580" Nov 26 06:40:08 crc kubenswrapper[4775]: E1126 06:40:08.033804 4775 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 26 06:40:08 crc kubenswrapper[4775]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xrngt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-tfrkk_openstack(ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 26 06:40:08 crc kubenswrapper[4775]: > logger="UnhandledError" Nov 26 06:40:08 crc kubenswrapper[4775]: E1126 06:40:08.034921 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.037420 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4qmsw"] Nov 26 06:40:08 crc kubenswrapper[4775]: W1126 06:40:08.104048 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64dca619_067a_4f6d_b391_c736d9771a1a.slice/crio-1b38ad46a8b3e7586732f72da8c1540fcdc70923d2d1de38262de000653f863d WatchSource:0}: Error finding container 1b38ad46a8b3e7586732f72da8c1540fcdc70923d2d1de38262de000653f863d: Status 404 returned error can't find the container with id 1b38ad46a8b3e7586732f72da8c1540fcdc70923d2d1de38262de000653f863d Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.645777 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e61da90a-7acf-4b95-934b-69f09efb5cbc","Type":"ContainerStarted","Data":"f2968b247ee9c882653976e48da49433a59b48206ec800867a9a102aa2d19214"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.648428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bf022b43-4760-4e21-8f5b-2558a69ede14","Type":"ContainerStarted","Data":"4d56ae7cd0b0fa5dd9e623a252ff9bbf56d8dc19ed117b411dbcc00a0334dfe8"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.650040 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"633ccd9d-4b8e-4797-89f9-7412c416b35b","Type":"ContainerStarted","Data":"71f48b6f463433d07630a6a6f21c8af9a1f079643e398ec289111cf0d362ec72"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.651544 4775 generic.go:334] "Generic (PLEG): container finished" podID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerID="f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2" exitCode=0 Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.651583 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerDied","Data":"f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.655028 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerID="42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637" exitCode=0 Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.655093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rmnn" event={"ID":"1c87f5c8-5d09-4e93-ad84-a2945745039b","Type":"ContainerDied","Data":"42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.658373 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" event={"ID":"01958f53-a2a9-4783-9669-abd002d57a66","Type":"ContainerStarted","Data":"a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.658470 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.660923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4mqj" event={"ID":"d30382f9-28bd-4ab7-8fec-1f8932b3f804","Type":"ContainerStarted","Data":"9c60d82c75f28a098d5b1b17a581d502a38b36a443f1667d0b9bc231203743a4"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.663078 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4qmsw" event={"ID":"64dca619-067a-4f6d-b391-c736d9771a1a","Type":"ContainerStarted","Data":"1b38ad46a8b3e7586732f72da8c1540fcdc70923d2d1de38262de000653f863d"} Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.690218 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" podStartSLOduration=16.690203241 podStartE2EDuration="16.690203241s" podCreationTimestamp="2025-11-26 06:39:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:40:08.684646871 +0000 UTC m=+892.045950843" watchObservedRunningTime="2025-11-26 06:40:08.690203241 +0000 UTC m=+892.051507193" Nov 26 06:40:08 crc kubenswrapper[4775]: I1126 06:40:08.849326 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 26 06:40:09 crc kubenswrapper[4775]: W1126 06:40:09.002514 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc88dc0e_14a1_479e_9682_bb5c611259a8.slice/crio-017eb8f524957c17b631ad24a1f55af934f3cb019d56202fcf368c30ff145aab WatchSource:0}: Error finding container 017eb8f524957c17b631ad24a1f55af934f3cb019d56202fcf368c30ff145aab: Status 404 returned error can't find the container with id 017eb8f524957c17b631ad24a1f55af934f3cb019d56202fcf368c30ff145aab Nov 26 06:40:09 crc kubenswrapper[4775]: I1126 06:40:09.343124 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6a7ce1d-2609-4d78-9565-c900247f40d6" path="/var/lib/kubelet/pods/d6a7ce1d-2609-4d78-9565-c900247f40d6/volumes" Nov 26 06:40:09 crc kubenswrapper[4775]: I1126 06:40:09.343738 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6381ad9-d93b-4f52-a5ae-4d694f4db96d" path="/var/lib/kubelet/pods/f6381ad9-d93b-4f52-a5ae-4d694f4db96d/volumes" Nov 26 06:40:09 crc kubenswrapper[4775]: I1126 06:40:09.422294 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:40:09 crc kubenswrapper[4775]: I1126 06:40:09.422461 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:40:09 crc kubenswrapper[4775]: I1126 06:40:09.671458 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc88dc0e-14a1-479e-9682-bb5c611259a8","Type":"ContainerStarted","Data":"017eb8f524957c17b631ad24a1f55af934f3cb019d56202fcf368c30ff145aab"} Nov 26 06:40:10 crc kubenswrapper[4775]: I1126 06:40:10.493580 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-258zm" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="registry-server" probeResult="failure" output=< Nov 26 06:40:10 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Nov 26 06:40:10 crc kubenswrapper[4775]: > Nov 26 06:40:12 crc kubenswrapper[4775]: I1126 06:40:12.794777 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:40:12 crc kubenswrapper[4775]: I1126 06:40:12.856201 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tfrkk"] Nov 26 06:40:17 crc kubenswrapper[4775]: I1126 06:40:17.748037 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" event={"ID":"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06","Type":"ContainerStarted","Data":"d3d8bf50cc4011580997dec0de0a58ddd41324f8d8e9ebe5e2345ec1b7ba292c"} Nov 26 06:40:17 crc kubenswrapper[4775]: I1126 06:40:17.748633 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:40:17 crc kubenswrapper[4775]: I1126 06:40:17.748200 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerName="dnsmasq-dns" containerID="cri-o://d3d8bf50cc4011580997dec0de0a58ddd41324f8d8e9ebe5e2345ec1b7ba292c" gracePeriod=10 Nov 26 06:40:17 crc kubenswrapper[4775]: I1126 06:40:17.751511 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerStarted","Data":"00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a"} Nov 26 06:40:17 crc kubenswrapper[4775]: I1126 06:40:17.771702 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" podStartSLOduration=23.973043001 podStartE2EDuration="25.771684049s" podCreationTimestamp="2025-11-26 06:39:52 +0000 UTC" firstStartedPulling="2025-11-26 06:40:05.191040625 +0000 UTC m=+888.552344587" lastFinishedPulling="2025-11-26 06:40:06.989681683 +0000 UTC m=+890.350985635" observedRunningTime="2025-11-26 06:40:17.766993858 +0000 UTC m=+901.128297810" watchObservedRunningTime="2025-11-26 06:40:17.771684049 +0000 UTC m=+901.132988001" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.762793 4775 generic.go:334] "Generic (PLEG): container finished" podID="64dca619-067a-4f6d-b391-c736d9771a1a" containerID="e3e74c9e73e8e6b4d19814d7fde107ab91bf205b7015a21e02d538dcf0f9c9a2" exitCode=0 Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.763024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4qmsw" event={"ID":"64dca619-067a-4f6d-b391-c736d9771a1a","Type":"ContainerDied","Data":"e3e74c9e73e8e6b4d19814d7fde107ab91bf205b7015a21e02d538dcf0f9c9a2"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.765350 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e61da90a-7acf-4b95-934b-69f09efb5cbc","Type":"ContainerStarted","Data":"719872f91c3b9b99b51855652285ae19924c1a85fe09e87396d52256b35ad47b"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.766695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"bf022b43-4760-4e21-8f5b-2558a69ede14","Type":"ContainerStarted","Data":"7b1b15abfe3cf1e84a704fe96858ed5834d3ae33daf401cd9d29e521e8f16797"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.767068 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.770218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0028de9f-a7db-47c0-a552-e9f518435458","Type":"ContainerStarted","Data":"8dc3516129abeb9abf29ccc070f8ca0703e878a0f4ceda1dac6d135e2f1978ac"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.772993 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc88dc0e-14a1-479e-9682-bb5c611259a8","Type":"ContainerStarted","Data":"3bca636696d1252edd4fbb564a5c9c9ef7bfa94a649f8823e8ff9196b965e380"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.775309 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerID="f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87" exitCode=0 Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.775379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rmnn" event={"ID":"1c87f5c8-5d09-4e93-ad84-a2945745039b","Type":"ContainerDied","Data":"f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.777425 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c","Type":"ContainerStarted","Data":"bb7149edccc98fa759bad7137fe2ba1de2ee046dcb524b3597987fa8f8ef1aed"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.780591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"633ccd9d-4b8e-4797-89f9-7412c416b35b","Type":"ContainerStarted","Data":"f9ef220083717002d290f22b3a464b503d762b539cfc24ac41155ce7b03aaef3"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.781139 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.784737 4775 generic.go:334] "Generic (PLEG): container finished" podID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerID="d3d8bf50cc4011580997dec0de0a58ddd41324f8d8e9ebe5e2345ec1b7ba292c" exitCode=0 Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.784811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" event={"ID":"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06","Type":"ContainerDied","Data":"d3d8bf50cc4011580997dec0de0a58ddd41324f8d8e9ebe5e2345ec1b7ba292c"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.784840 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" event={"ID":"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06","Type":"ContainerDied","Data":"e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.784854 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e367567d938059df60693d9b9650a04c6d8b9a895f448e052a77bb12d1efa580" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.790440 4775 generic.go:334] "Generic (PLEG): container finished" podID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerID="00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a" exitCode=0 Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.790721 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerDied","Data":"00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.793133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4mqj" event={"ID":"d30382f9-28bd-4ab7-8fec-1f8932b3f804","Type":"ContainerStarted","Data":"433bf7e600b0e442f420fecc3bd4ad86adfd5a801cd38dd64c9441f29364d2ba"} Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.793896 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.818143 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.425069279 podStartE2EDuration="22.818123314s" podCreationTimestamp="2025-11-26 06:39:56 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.57000238 +0000 UTC m=+890.931306342" lastFinishedPulling="2025-11-26 06:40:15.963056415 +0000 UTC m=+899.324360377" observedRunningTime="2025-11-26 06:40:18.817739435 +0000 UTC m=+902.179043387" watchObservedRunningTime="2025-11-26 06:40:18.818123314 +0000 UTC m=+902.179427266" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.829641 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.321764555 podStartE2EDuration="20.829621795s" podCreationTimestamp="2025-11-26 06:39:58 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.883877877 +0000 UTC m=+891.245181829" lastFinishedPulling="2025-11-26 06:40:17.391735107 +0000 UTC m=+900.753039069" observedRunningTime="2025-11-26 06:40:18.829311768 +0000 UTC m=+902.190615720" watchObservedRunningTime="2025-11-26 06:40:18.829621795 +0000 UTC m=+902.190925737" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.903198 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-k4mqj" podStartSLOduration=8.160592276 podStartE2EDuration="16.903182806s" podCreationTimestamp="2025-11-26 06:40:02 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.893550535 +0000 UTC m=+891.254854487" lastFinishedPulling="2025-11-26 06:40:16.636141055 +0000 UTC m=+899.997445017" observedRunningTime="2025-11-26 06:40:18.897347939 +0000 UTC m=+902.258651891" watchObservedRunningTime="2025-11-26 06:40:18.903182806 +0000 UTC m=+902.264486758" Nov 26 06:40:18 crc kubenswrapper[4775]: I1126 06:40:18.987335 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.167026 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrngt\" (UniqueName: \"kubernetes.io/projected/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-kube-api-access-xrngt\") pod \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.167145 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-config\") pod \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.167173 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc\") pod \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.174207 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-kube-api-access-xrngt" (OuterVolumeSpecName: "kube-api-access-xrngt") pod "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" (UID: "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06"). InnerVolumeSpecName "kube-api-access-xrngt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:19 crc kubenswrapper[4775]: E1126 06:40:19.205683 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc podName:ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:19.705657774 +0000 UTC m=+903.066961726 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc") pod "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" (UID: "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06") : error deleting /var/lib/kubelet/pods/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06/volume-subpaths: remove /var/lib/kubelet/pods/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06/volume-subpaths: no such file or directory Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.206002 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-config" (OuterVolumeSpecName: "config") pod "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" (UID: "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.269029 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrngt\" (UniqueName: \"kubernetes.io/projected/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-kube-api-access-xrngt\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.269061 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.572606 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.614934 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.777148 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc\") pod \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\" (UID: \"ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06\") " Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.778537 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" (UID: "ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.803524 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4qmsw" event={"ID":"64dca619-067a-4f6d-b391-c736d9771a1a","Type":"ContainerStarted","Data":"b75010bf4608275b676ed7363f883cd52e5a921e2a7c0f729e2d5c5fb25ded64"} Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.803563 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4qmsw" event={"ID":"64dca619-067a-4f6d-b391-c736d9771a1a","Type":"ContainerStarted","Data":"8cd2deaded5c801913ae677ae249b0f82b62c0861fa6df8e3d74fcccf90dbf62"} Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.804073 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.804382 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.811020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"92a5651e-7175-45ca-a69b-dc19a11f2943","Type":"ContainerStarted","Data":"bcdd2a68e5172c990298b42de9bed69bd8c00801bcd6c8e980fd5d7cc6fafff6"} Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.812435 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63f863dc-36a4-4325-b520-e26b3dca309e","Type":"ContainerStarted","Data":"986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf"} Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.814390 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerStarted","Data":"1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5"} Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.816490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rmnn" event={"ID":"1c87f5c8-5d09-4e93-ad84-a2945745039b","Type":"ContainerStarted","Data":"96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1"} Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.816605 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-tfrkk" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.843090 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-4qmsw" podStartSLOduration=9.48437438 podStartE2EDuration="17.843065796s" podCreationTimestamp="2025-11-26 06:40:02 +0000 UTC" firstStartedPulling="2025-11-26 06:40:08.106378253 +0000 UTC m=+891.467682205" lastFinishedPulling="2025-11-26 06:40:16.465069659 +0000 UTC m=+899.826373621" observedRunningTime="2025-11-26 06:40:19.833509369 +0000 UTC m=+903.194813341" watchObservedRunningTime="2025-11-26 06:40:19.843065796 +0000 UTC m=+903.204369748" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.878970 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.910360 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9rmnn" podStartSLOduration=4.383865081 podStartE2EDuration="14.910341877s" podCreationTimestamp="2025-11-26 06:40:05 +0000 UTC" firstStartedPulling="2025-11-26 06:40:08.656823016 +0000 UTC m=+892.018126968" lastFinishedPulling="2025-11-26 06:40:19.183299812 +0000 UTC m=+902.544603764" observedRunningTime="2025-11-26 06:40:19.895800466 +0000 UTC m=+903.257104428" watchObservedRunningTime="2025-11-26 06:40:19.910341877 +0000 UTC m=+903.271645849" Nov 26 06:40:19 crc kubenswrapper[4775]: I1126 06:40:19.970764 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tfrkk"] Nov 26 06:40:20 crc kubenswrapper[4775]: I1126 06:40:20.002978 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-tfrkk"] Nov 26 06:40:20 crc kubenswrapper[4775]: I1126 06:40:20.046656 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-258zm"] Nov 26 06:40:20 crc kubenswrapper[4775]: I1126 06:40:20.058178 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hb7d2" podStartSLOduration=7.522196278 podStartE2EDuration="18.058154167s" podCreationTimestamp="2025-11-26 06:40:02 +0000 UTC" firstStartedPulling="2025-11-26 06:40:08.653221061 +0000 UTC m=+892.014525013" lastFinishedPulling="2025-11-26 06:40:19.18917895 +0000 UTC m=+902.550482902" observedRunningTime="2025-11-26 06:40:20.02447054 +0000 UTC m=+903.385774492" watchObservedRunningTime="2025-11-26 06:40:20.058154167 +0000 UTC m=+903.419458129" Nov 26 06:40:20 crc kubenswrapper[4775]: I1126 06:40:20.825617 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-258zm" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="registry-server" containerID="cri-o://f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412" gracePeriod=2 Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.347344 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" path="/var/lib/kubelet/pods/ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06/volumes" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.420015 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.420074 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.428660 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.526141 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-utilities\") pod \"193ae057-ca33-4654-95f8-5453f0938ce7\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.526220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-catalog-content\") pod \"193ae057-ca33-4654-95f8-5453f0938ce7\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.526346 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k2sl\" (UniqueName: \"kubernetes.io/projected/193ae057-ca33-4654-95f8-5453f0938ce7-kube-api-access-2k2sl\") pod \"193ae057-ca33-4654-95f8-5453f0938ce7\" (UID: \"193ae057-ca33-4654-95f8-5453f0938ce7\") " Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.527325 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-utilities" (OuterVolumeSpecName: "utilities") pod "193ae057-ca33-4654-95f8-5453f0938ce7" (UID: "193ae057-ca33-4654-95f8-5453f0938ce7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.528054 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.531402 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193ae057-ca33-4654-95f8-5453f0938ce7-kube-api-access-2k2sl" (OuterVolumeSpecName: "kube-api-access-2k2sl") pod "193ae057-ca33-4654-95f8-5453f0938ce7" (UID: "193ae057-ca33-4654-95f8-5453f0938ce7"). InnerVolumeSpecName "kube-api-access-2k2sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.626244 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "193ae057-ca33-4654-95f8-5453f0938ce7" (UID: "193ae057-ca33-4654-95f8-5453f0938ce7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.629461 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k2sl\" (UniqueName: \"kubernetes.io/projected/193ae057-ca33-4654-95f8-5453f0938ce7-kube-api-access-2k2sl\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.629499 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/193ae057-ca33-4654-95f8-5453f0938ce7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.832890 4775 generic.go:334] "Generic (PLEG): container finished" podID="0028de9f-a7db-47c0-a552-e9f518435458" containerID="8dc3516129abeb9abf29ccc070f8ca0703e878a0f4ceda1dac6d135e2f1978ac" exitCode=0 Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.833065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0028de9f-a7db-47c0-a552-e9f518435458","Type":"ContainerDied","Data":"8dc3516129abeb9abf29ccc070f8ca0703e878a0f4ceda1dac6d135e2f1978ac"} Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.839661 4775 generic.go:334] "Generic (PLEG): container finished" podID="193ae057-ca33-4654-95f8-5453f0938ce7" containerID="f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412" exitCode=0 Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.839803 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-258zm" event={"ID":"193ae057-ca33-4654-95f8-5453f0938ce7","Type":"ContainerDied","Data":"f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412"} Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.839844 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-258zm" event={"ID":"193ae057-ca33-4654-95f8-5453f0938ce7","Type":"ContainerDied","Data":"2d9685ad860baca1304daaedf298ad35fbdde32e9a306336b7ff2162db30c5d3"} Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.839875 4775 scope.go:117] "RemoveContainer" containerID="f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.840047 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-258zm" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.843698 4775 generic.go:334] "Generic (PLEG): container finished" podID="7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c" containerID="bb7149edccc98fa759bad7137fe2ba1de2ee046dcb524b3597987fa8f8ef1aed" exitCode=0 Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.843746 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c","Type":"ContainerDied","Data":"bb7149edccc98fa759bad7137fe2ba1de2ee046dcb524b3597987fa8f8ef1aed"} Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.866073 4775 scope.go:117] "RemoveContainer" containerID="fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.886620 4775 scope.go:117] "RemoveContainer" containerID="446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.904877 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-258zm"] Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.909984 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-258zm"] Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.931327 4775 scope.go:117] "RemoveContainer" containerID="f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412" Nov 26 06:40:21 crc kubenswrapper[4775]: E1126 06:40:21.931761 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412\": container with ID starting with f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412 not found: ID does not exist" containerID="f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.931822 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412"} err="failed to get container status \"f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412\": rpc error: code = NotFound desc = could not find container \"f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412\": container with ID starting with f636808a996b52890b753bc6872f5c7f9348316747ca28e0dc65cc0557a8c412 not found: ID does not exist" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.931863 4775 scope.go:117] "RemoveContainer" containerID="fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1" Nov 26 06:40:21 crc kubenswrapper[4775]: E1126 06:40:21.932533 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1\": container with ID starting with fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1 not found: ID does not exist" containerID="fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.932557 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1"} err="failed to get container status \"fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1\": rpc error: code = NotFound desc = could not find container \"fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1\": container with ID starting with fe1f8872836493f7e15b7f74622c10ca7d235fa27de9cba3d25babbe73d7fea1 not found: ID does not exist" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.932570 4775 scope.go:117] "RemoveContainer" containerID="446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945" Nov 26 06:40:21 crc kubenswrapper[4775]: E1126 06:40:21.932971 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945\": container with ID starting with 446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945 not found: ID does not exist" containerID="446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945" Nov 26 06:40:21 crc kubenswrapper[4775]: I1126 06:40:21.932990 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945"} err="failed to get container status \"446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945\": rpc error: code = NotFound desc = could not find container \"446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945\": container with ID starting with 446f540f5753def9d118741ad85578a5ef9df225642330d979621ca357cb2945 not found: ID does not exist" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.174564 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.404334 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.404374 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.441155 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.857871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0028de9f-a7db-47c0-a552-e9f518435458","Type":"ContainerStarted","Data":"f0076df01fed37e56541f8b5d64f069e94585e56df327e4ca8566b6c09598c99"} Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.861102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dc88dc0e-14a1-479e-9682-bb5c611259a8","Type":"ContainerStarted","Data":"e07fd8f5447952533def79bbb14f032992eec01bd47de22ed206bc664a48d60d"} Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.865383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c","Type":"ContainerStarted","Data":"d3c30456a6878272a6e4b777b81d6fe67f82c7bd4be68e34295c250a4adc8f80"} Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.868063 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"e61da90a-7acf-4b95-934b-69f09efb5cbc","Type":"ContainerStarted","Data":"4a12181441f558ef11b461e32cbe943e253a8870d459f407876792ac3c844dc4"} Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.898537 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.635648908 podStartE2EDuration="29.898521703s" podCreationTimestamp="2025-11-26 06:39:53 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.10980864 +0000 UTC m=+890.471112592" lastFinishedPulling="2025-11-26 06:40:16.372681435 +0000 UTC m=+899.733985387" observedRunningTime="2025-11-26 06:40:22.895187213 +0000 UTC m=+906.256491165" watchObservedRunningTime="2025-11-26 06:40:22.898521703 +0000 UTC m=+906.259825655" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.938656 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.923043147 podStartE2EDuration="27.938626362s" podCreationTimestamp="2025-11-26 06:39:55 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.540380723 +0000 UTC m=+890.901684675" lastFinishedPulling="2025-11-26 06:40:16.555963898 +0000 UTC m=+899.917267890" observedRunningTime="2025-11-26 06:40:22.926195898 +0000 UTC m=+906.287499870" watchObservedRunningTime="2025-11-26 06:40:22.938626362 +0000 UTC m=+906.299930354" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.949542 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.272807966 podStartE2EDuration="17.949522096s" podCreationTimestamp="2025-11-26 06:40:05 +0000 UTC" firstStartedPulling="2025-11-26 06:40:09.004488538 +0000 UTC m=+892.365792490" lastFinishedPulling="2025-11-26 06:40:21.681202668 +0000 UTC m=+905.042506620" observedRunningTime="2025-11-26 06:40:22.94783091 +0000 UTC m=+906.309134872" watchObservedRunningTime="2025-11-26 06:40:22.949522096 +0000 UTC m=+906.310826058" Nov 26 06:40:22 crc kubenswrapper[4775]: I1126 06:40:22.977335 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.270739786 podStartE2EDuration="21.977315844s" podCreationTimestamp="2025-11-26 06:40:01 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.966819719 +0000 UTC m=+891.328123671" lastFinishedPulling="2025-11-26 06:40:21.673395777 +0000 UTC m=+905.034699729" observedRunningTime="2025-11-26 06:40:22.971445166 +0000 UTC m=+906.332749138" watchObservedRunningTime="2025-11-26 06:40:22.977315844 +0000 UTC m=+906.338619806" Nov 26 06:40:23 crc kubenswrapper[4775]: I1126 06:40:23.340018 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" path="/var/lib/kubelet/pods/193ae057-ca33-4654-95f8-5453f0938ce7/volumes" Nov 26 06:40:23 crc kubenswrapper[4775]: I1126 06:40:23.341605 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.338224 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.413683 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.446451 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.515287 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.900502 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.962141 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 26 06:40:24 crc kubenswrapper[4775]: I1126 06:40:24.974502 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.253602 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gzwzp"] Nov 26 06:40:25 crc kubenswrapper[4775]: E1126 06:40:25.253985 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerName="init" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254002 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerName="init" Nov 26 06:40:25 crc kubenswrapper[4775]: E1126 06:40:25.254012 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerName="dnsmasq-dns" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254018 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerName="dnsmasq-dns" Nov 26 06:40:25 crc kubenswrapper[4775]: E1126 06:40:25.254033 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="extract-content" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254039 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="extract-content" Nov 26 06:40:25 crc kubenswrapper[4775]: E1126 06:40:25.254076 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="registry-server" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254083 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="registry-server" Nov 26 06:40:25 crc kubenswrapper[4775]: E1126 06:40:25.254099 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="extract-utilities" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254106 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="extract-utilities" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254252 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="193ae057-ca33-4654-95f8-5453f0938ce7" containerName="registry-server" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.254265 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceab34dd-6af2-44c4-8e4d-9ad3a92b3c06" containerName="dnsmasq-dns" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.255182 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.259397 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.268763 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gzwzp"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.294422 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-kd448"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.295364 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: W1126 06:40:25.305785 4775 reflector.go:561] object-"openstack"/"ovncontroller-metrics-config": failed to list *v1.ConfigMap: configmaps "ovncontroller-metrics-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Nov 26 06:40:25 crc kubenswrapper[4775]: E1126 06:40:25.305846 4775 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovncontroller-metrics-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovncontroller-metrics-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.376887 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kd448"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416588 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz6sw\" (UniqueName: \"kubernetes.io/projected/ff2f6960-e698-4a01-9a6d-9b1fd470e295-kube-api-access-gz6sw\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416671 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-config\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416707 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff2f6960-e698-4a01-9a6d-9b1fd470e295-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416756 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff2f6960-e698-4a01-9a6d-9b1fd470e295-config\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416811 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416844 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ff2f6960-e698-4a01-9a6d-9b1fd470e295-ovn-rundir\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416908 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416942 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ff2f6960-e698-4a01-9a6d-9b1fd470e295-ovs-rundir\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.416976 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vp65\" (UniqueName: \"kubernetes.io/projected/fa323792-e445-455b-bbaf-fc62e6882888-kube-api-access-5vp65\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.417010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff2f6960-e698-4a01-9a6d-9b1fd470e295-combined-ca-bundle\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.427397 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.427457 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.494334 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.495608 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.499745 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.499999 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.502439 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.502683 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-z2bp8" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.518671 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.518748 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ff2f6960-e698-4a01-9a6d-9b1fd470e295-ovn-rundir\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.518817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.518877 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ff2f6960-e698-4a01-9a6d-9b1fd470e295-ovs-rundir\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.518904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vp65\" (UniqueName: \"kubernetes.io/projected/fa323792-e445-455b-bbaf-fc62e6882888-kube-api-access-5vp65\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519126 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ff2f6960-e698-4a01-9a6d-9b1fd470e295-ovn-rundir\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519257 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff2f6960-e698-4a01-9a6d-9b1fd470e295-combined-ca-bundle\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519331 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ff2f6960-e698-4a01-9a6d-9b1fd470e295-ovs-rundir\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz6sw\" (UniqueName: \"kubernetes.io/projected/ff2f6960-e698-4a01-9a6d-9b1fd470e295-kube-api-access-gz6sw\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519419 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-config\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff2f6960-e698-4a01-9a6d-9b1fd470e295-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519504 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff2f6960-e698-4a01-9a6d-9b1fd470e295-config\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.519588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.520784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.520976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-config\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.534112 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff2f6960-e698-4a01-9a6d-9b1fd470e295-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.542270 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz6sw\" (UniqueName: \"kubernetes.io/projected/ff2f6960-e698-4a01-9a6d-9b1fd470e295-kube-api-access-gz6sw\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.547470 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vp65\" (UniqueName: \"kubernetes.io/projected/fa323792-e445-455b-bbaf-fc62e6882888-kube-api-access-5vp65\") pod \"dnsmasq-dns-7fd796d7df-gzwzp\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.578123 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.588750 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.596439 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff2f6960-e698-4a01-9a6d-9b1fd470e295-combined-ca-bundle\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621174 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n666\" (UniqueName: \"kubernetes.io/projected/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-kube-api-access-8n666\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621241 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-scripts\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621265 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-config\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.621381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.629692 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gzwzp"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.639349 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mb28z"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.640826 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.645151 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mb28z"] Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.650606 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.722889 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.722937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723013 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723051 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723097 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n666\" (UniqueName: \"kubernetes.io/projected/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-kube-api-access-8n666\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723119 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-scripts\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723187 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-config\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723239 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-config\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.723328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2f7p\" (UniqueName: \"kubernetes.io/projected/b2dd29bc-afa3-445a-be94-1b031b706b3d-kube-api-access-b2f7p\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.726310 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-scripts\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.726393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.726470 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.726496 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.726967 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-config\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.736121 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.751746 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n666\" (UniqueName: \"kubernetes.io/projected/c2f5d800-59f5-41fa-b45b-f2fa0981b3da-kube-api-access-8n666\") pod \"ovn-northd-0\" (UID: \"c2f5d800-59f5-41fa-b45b-f2fa0981b3da\") " pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.792143 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.792233 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.812344 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.824938 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.825292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-config\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.825323 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2f7p\" (UniqueName: \"kubernetes.io/projected/b2dd29bc-afa3-445a-be94-1b031b706b3d-kube-api-access-b2f7p\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.825358 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.825412 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.825847 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.826233 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-config\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.826407 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.827555 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.846603 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2f7p\" (UniqueName: \"kubernetes.io/projected/b2dd29bc-afa3-445a-be94-1b031b706b3d-kube-api-access-b2f7p\") pod \"dnsmasq-dns-86db49b7ff-mb28z\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.850508 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.955146 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:25 crc kubenswrapper[4775]: I1126 06:40:25.993210 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.053546 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gzwzp"] Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.224152 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9rmnn"] Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.257515 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 26 06:40:26 crc kubenswrapper[4775]: W1126 06:40:26.261400 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2f5d800_59f5_41fa_b45b_f2fa0981b3da.slice/crio-0b59d9a39d4d4e9f68434ee274982327fdf9273f5ef544068e277ff9d5fffe54 WatchSource:0}: Error finding container 0b59d9a39d4d4e9f68434ee274982327fdf9273f5ef544068e277ff9d5fffe54: Status 404 returned error can't find the container with id 0b59d9a39d4d4e9f68434ee274982327fdf9273f5ef544068e277ff9d5fffe54 Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.421013 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mb28z"] Nov 26 06:40:26 crc kubenswrapper[4775]: W1126 06:40:26.423835 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2dd29bc_afa3_445a_be94_1b031b706b3d.slice/crio-28f26e97931f9f820ad69dccc60c7881058156685c04bfa040d17c18807c83b2 WatchSource:0}: Error finding container 28f26e97931f9f820ad69dccc60c7881058156685c04bfa040d17c18807c83b2: Status 404 returned error can't find the container with id 28f26e97931f9f820ad69dccc60c7881058156685c04bfa040d17c18807c83b2 Nov 26 06:40:26 crc kubenswrapper[4775]: E1126 06:40:26.521121 4775 configmap.go:193] Couldn't get configMap openstack/ovncontroller-metrics-config: failed to sync configmap cache: timed out waiting for the condition Nov 26 06:40:26 crc kubenswrapper[4775]: E1126 06:40:26.521216 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ff2f6960-e698-4a01-9a6d-9b1fd470e295-config podName:ff2f6960-e698-4a01-9a6d-9b1fd470e295 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:27.021192972 +0000 UTC m=+910.382496944 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/ff2f6960-e698-4a01-9a6d-9b1fd470e295-config") pod "ovn-controller-metrics-kd448" (UID: "ff2f6960-e698-4a01-9a6d-9b1fd470e295") : failed to sync configmap cache: timed out waiting for the condition Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.602305 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.742258 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.742730 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.918800 4775 generic.go:334] "Generic (PLEG): container finished" podID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerID="9aca70db95c61976ec99d30423b7c9711fc91d7236748ac127bce2afc2fa69f5" exitCode=0 Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.918935 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" event={"ID":"b2dd29bc-afa3-445a-be94-1b031b706b3d","Type":"ContainerDied","Data":"9aca70db95c61976ec99d30423b7c9711fc91d7236748ac127bce2afc2fa69f5"} Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.918976 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" event={"ID":"b2dd29bc-afa3-445a-be94-1b031b706b3d","Type":"ContainerStarted","Data":"28f26e97931f9f820ad69dccc60c7881058156685c04bfa040d17c18807c83b2"} Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.921024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2f5d800-59f5-41fa-b45b-f2fa0981b3da","Type":"ContainerStarted","Data":"0b59d9a39d4d4e9f68434ee274982327fdf9273f5ef544068e277ff9d5fffe54"} Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.928814 4775 generic.go:334] "Generic (PLEG): container finished" podID="fa323792-e445-455b-bbaf-fc62e6882888" containerID="a54bf2a1a83cc30e36038c6fca15065d3a284237deb899b1a394f8755b2b35a9" exitCode=0 Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.929681 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" event={"ID":"fa323792-e445-455b-bbaf-fc62e6882888","Type":"ContainerDied","Data":"a54bf2a1a83cc30e36038c6fca15065d3a284237deb899b1a394f8755b2b35a9"} Nov 26 06:40:26 crc kubenswrapper[4775]: I1126 06:40:26.929707 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" event={"ID":"fa323792-e445-455b-bbaf-fc62e6882888","Type":"ContainerStarted","Data":"6153bac7d5faea2d1a538f0735c59023af5d17e9c8bf438605984f2ad3a4454e"} Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.050858 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff2f6960-e698-4a01-9a6d-9b1fd470e295-config\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.056103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff2f6960-e698-4a01-9a6d-9b1fd470e295-config\") pod \"ovn-controller-metrics-kd448\" (UID: \"ff2f6960-e698-4a01-9a6d-9b1fd470e295\") " pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.115578 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-kd448" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.251161 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.354589 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-dns-svc\") pod \"fa323792-e445-455b-bbaf-fc62e6882888\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.354629 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-ovsdbserver-nb\") pod \"fa323792-e445-455b-bbaf-fc62e6882888\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.354649 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vp65\" (UniqueName: \"kubernetes.io/projected/fa323792-e445-455b-bbaf-fc62e6882888-kube-api-access-5vp65\") pod \"fa323792-e445-455b-bbaf-fc62e6882888\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.354680 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-config\") pod \"fa323792-e445-455b-bbaf-fc62e6882888\" (UID: \"fa323792-e445-455b-bbaf-fc62e6882888\") " Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.382168 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa323792-e445-455b-bbaf-fc62e6882888-kube-api-access-5vp65" (OuterVolumeSpecName: "kube-api-access-5vp65") pod "fa323792-e445-455b-bbaf-fc62e6882888" (UID: "fa323792-e445-455b-bbaf-fc62e6882888"). InnerVolumeSpecName "kube-api-access-5vp65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.389410 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fa323792-e445-455b-bbaf-fc62e6882888" (UID: "fa323792-e445-455b-bbaf-fc62e6882888"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.410341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-config" (OuterVolumeSpecName: "config") pod "fa323792-e445-455b-bbaf-fc62e6882888" (UID: "fa323792-e445-455b-bbaf-fc62e6882888"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.452215 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fa323792-e445-455b-bbaf-fc62e6882888" (UID: "fa323792-e445-455b-bbaf-fc62e6882888"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.464775 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.464806 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.464816 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vp65\" (UniqueName: \"kubernetes.io/projected/fa323792-e445-455b-bbaf-fc62e6882888-kube-api-access-5vp65\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.464824 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa323792-e445-455b-bbaf-fc62e6882888-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.573881 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-kd448"] Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.938379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kd448" event={"ID":"ff2f6960-e698-4a01-9a6d-9b1fd470e295","Type":"ContainerStarted","Data":"df3e221c8d651910f296aba26e62f9e4c7cec234e6541cba503166e33266d5e1"} Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.941164 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" event={"ID":"b2dd29bc-afa3-445a-be94-1b031b706b3d","Type":"ContainerStarted","Data":"0bfb6ef37bc33f00ae0701eda6eec33eb9de2c8db1ae6685bda1710d3536364e"} Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.941326 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.954026 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" event={"ID":"fa323792-e445-455b-bbaf-fc62e6882888","Type":"ContainerDied","Data":"6153bac7d5faea2d1a538f0735c59023af5d17e9c8bf438605984f2ad3a4454e"} Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.954102 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-gzwzp" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.954104 4775 scope.go:117] "RemoveContainer" containerID="a54bf2a1a83cc30e36038c6fca15065d3a284237deb899b1a394f8755b2b35a9" Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.954085 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9rmnn" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="registry-server" containerID="cri-o://96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1" gracePeriod=2 Nov 26 06:40:27 crc kubenswrapper[4775]: I1126 06:40:27.970272 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" podStartSLOduration=2.970254047 podStartE2EDuration="2.970254047s" podCreationTimestamp="2025-11-26 06:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:40:27.9621772 +0000 UTC m=+911.323481162" watchObservedRunningTime="2025-11-26 06:40:27.970254047 +0000 UTC m=+911.331558009" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.030788 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gzwzp"] Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.038581 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-gzwzp"] Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.336438 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.381826 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-catalog-content\") pod \"1c87f5c8-5d09-4e93-ad84-a2945745039b\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.381875 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-utilities\") pod \"1c87f5c8-5d09-4e93-ad84-a2945745039b\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.381942 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g96km\" (UniqueName: \"kubernetes.io/projected/1c87f5c8-5d09-4e93-ad84-a2945745039b-kube-api-access-g96km\") pod \"1c87f5c8-5d09-4e93-ad84-a2945745039b\" (UID: \"1c87f5c8-5d09-4e93-ad84-a2945745039b\") " Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.382996 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-utilities" (OuterVolumeSpecName: "utilities") pod "1c87f5c8-5d09-4e93-ad84-a2945745039b" (UID: "1c87f5c8-5d09-4e93-ad84-a2945745039b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.419679 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c87f5c8-5d09-4e93-ad84-a2945745039b-kube-api-access-g96km" (OuterVolumeSpecName: "kube-api-access-g96km") pod "1c87f5c8-5d09-4e93-ad84-a2945745039b" (UID: "1c87f5c8-5d09-4e93-ad84-a2945745039b"). InnerVolumeSpecName "kube-api-access-g96km". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.483932 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.484957 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g96km\" (UniqueName: \"kubernetes.io/projected/1c87f5c8-5d09-4e93-ad84-a2945745039b-kube-api-access-g96km\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.610055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c87f5c8-5d09-4e93-ad84-a2945745039b" (UID: "1c87f5c8-5d09-4e93-ad84-a2945745039b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.688586 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c87f5c8-5d09-4e93-ad84-a2945745039b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.980677 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-kd448" event={"ID":"ff2f6960-e698-4a01-9a6d-9b1fd470e295","Type":"ContainerStarted","Data":"8018d794aea6fd8f4d1fd47ab9bc2a00aa1e6da603f6b69ea9e9db90e5895009"} Nov 26 06:40:28 crc kubenswrapper[4775]: I1126 06:40:28.993987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2f5d800-59f5-41fa-b45b-f2fa0981b3da","Type":"ContainerStarted","Data":"3041bcd8faf9104dba2578cd8f5c1433f8ec2220d099a9cc73c97cb80f491969"} Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.003397 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mb28z"] Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.019329 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-kd448" podStartSLOduration=4.019312163 podStartE2EDuration="4.019312163s" podCreationTimestamp="2025-11-26 06:40:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:40:29.01586899 +0000 UTC m=+912.377172942" watchObservedRunningTime="2025-11-26 06:40:29.019312163 +0000 UTC m=+912.380616115" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.028684 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerID="96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1" exitCode=0 Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.029120 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9rmnn" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.029763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rmnn" event={"ID":"1c87f5c8-5d09-4e93-ad84-a2945745039b","Type":"ContainerDied","Data":"96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1"} Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.029797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9rmnn" event={"ID":"1c87f5c8-5d09-4e93-ad84-a2945745039b","Type":"ContainerDied","Data":"a802c97b1c269018bdf1b9287cfebec903a62a18f5fcd854dfc1a424230657ed"} Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.029822 4775 scope.go:117] "RemoveContainer" containerID="96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.031912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.056647 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-2nsvj"] Nov 26 06:40:29 crc kubenswrapper[4775]: E1126 06:40:29.067909 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa323792-e445-455b-bbaf-fc62e6882888" containerName="init" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.067947 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa323792-e445-455b-bbaf-fc62e6882888" containerName="init" Nov 26 06:40:29 crc kubenswrapper[4775]: E1126 06:40:29.067961 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="extract-utilities" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.067968 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="extract-utilities" Nov 26 06:40:29 crc kubenswrapper[4775]: E1126 06:40:29.067985 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="extract-content" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.067993 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="extract-content" Nov 26 06:40:29 crc kubenswrapper[4775]: E1126 06:40:29.068004 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="registry-server" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.068010 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="registry-server" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.068221 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" containerName="registry-server" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.068236 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa323792-e445-455b-bbaf-fc62e6882888" containerName="init" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.069087 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2nsvj"] Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.069181 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.092298 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9rmnn"] Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.115311 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9rmnn"] Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.217568 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpqtf\" (UniqueName: \"kubernetes.io/projected/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-kube-api-access-cpqtf\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.217616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.217693 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-dns-svc\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.217741 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.217763 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-config\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.319075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpqtf\" (UniqueName: \"kubernetes.io/projected/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-kube-api-access-cpqtf\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.319162 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.319242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-dns-svc\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.319284 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.319312 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-config\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.320247 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-config\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.320359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.320388 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-dns-svc\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.321009 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.336816 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c87f5c8-5d09-4e93-ad84-a2945745039b" path="/var/lib/kubelet/pods/1c87f5c8-5d09-4e93-ad84-a2945745039b/volumes" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.337405 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa323792-e445-455b-bbaf-fc62e6882888" path="/var/lib/kubelet/pods/fa323792-e445-455b-bbaf-fc62e6882888/volumes" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.344792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpqtf\" (UniqueName: \"kubernetes.io/projected/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-kube-api-access-cpqtf\") pod \"dnsmasq-dns-698758b865-2nsvj\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:29 crc kubenswrapper[4775]: I1126 06:40:29.389707 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.040964 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c2f5d800-59f5-41fa-b45b-f2fa0981b3da","Type":"ContainerStarted","Data":"64fc2a0d0996a54fdf635eed9ef6b0d9b191598a118f67d60601cbb8105aad78"} Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.041127 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerName="dnsmasq-dns" containerID="cri-o://0bfb6ef37bc33f00ae0701eda6eec33eb9de2c8db1ae6685bda1710d3536364e" gracePeriod=10 Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.163596 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.169465 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.172781 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-np54r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.173012 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.175233 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.176134 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.178437 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.232645 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/38df1b80-132c-467f-969c-d3798dd1f7d9-cache\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.232694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.232813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.232855 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4klsv\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-kube-api-access-4klsv\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.232918 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/38df1b80-132c-467f-969c-d3798dd1f7d9-lock\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.334556 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.334630 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4klsv\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-kube-api-access-4klsv\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.334694 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/38df1b80-132c-467f-969c-d3798dd1f7d9-lock\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.334764 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/38df1b80-132c-467f-969c-d3798dd1f7d9-cache\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.334803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.335288 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.335431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/38df1b80-132c-467f-969c-d3798dd1f7d9-lock\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: E1126 06:40:30.334811 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 06:40:30 crc kubenswrapper[4775]: E1126 06:40:30.335507 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 06:40:30 crc kubenswrapper[4775]: E1126 06:40:30.335575 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift podName:38df1b80-132c-467f-969c-d3798dd1f7d9 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:30.835552142 +0000 UTC m=+914.196856194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift") pod "swift-storage-0" (UID: "38df1b80-132c-467f-969c-d3798dd1f7d9") : configmap "swift-ring-files" not found Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.335641 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/38df1b80-132c-467f-969c-d3798dd1f7d9-cache\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.361804 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4klsv\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-kube-api-access-4klsv\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.362970 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.439110 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ldf4r"] Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.440912 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.447865 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldf4r"] Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.537458 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-catalog-content\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.537828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-utilities\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.537942 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pftv\" (UniqueName: \"kubernetes.io/projected/0e3c05a3-6295-4d99-803a-a4495d2046fc-kube-api-access-8pftv\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.639057 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-utilities\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.639121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pftv\" (UniqueName: \"kubernetes.io/projected/0e3c05a3-6295-4d99-803a-a4495d2046fc-kube-api-access-8pftv\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.639172 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-catalog-content\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.639709 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-utilities\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.640914 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-catalog-content\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.657397 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pftv\" (UniqueName: \"kubernetes.io/projected/0e3c05a3-6295-4d99-803a-a4495d2046fc-kube-api-access-8pftv\") pod \"redhat-marketplace-ldf4r\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.807542 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:30 crc kubenswrapper[4775]: I1126 06:40:30.842506 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:30 crc kubenswrapper[4775]: E1126 06:40:30.842709 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 06:40:30 crc kubenswrapper[4775]: E1126 06:40:30.842959 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 06:40:30 crc kubenswrapper[4775]: E1126 06:40:30.843021 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift podName:38df1b80-132c-467f-969c-d3798dd1f7d9 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:31.843002055 +0000 UTC m=+915.204306007 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift") pod "swift-storage-0" (UID: "38df1b80-132c-467f-969c-d3798dd1f7d9") : configmap "swift-ring-files" not found Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.052231 4775 generic.go:334] "Generic (PLEG): container finished" podID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerID="0bfb6ef37bc33f00ae0701eda6eec33eb9de2c8db1ae6685bda1710d3536364e" exitCode=0 Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.052308 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" event={"ID":"b2dd29bc-afa3-445a-be94-1b031b706b3d","Type":"ContainerDied","Data":"0bfb6ef37bc33f00ae0701eda6eec33eb9de2c8db1ae6685bda1710d3536364e"} Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.171110 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.172162 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.268240 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.282823 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.401879 4775 scope.go:117] "RemoveContainer" containerID="f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.552914 4775 scope.go:117] "RemoveContainer" containerID="42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.572610 4775 scope.go:117] "RemoveContainer" containerID="96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1" Nov 26 06:40:31 crc kubenswrapper[4775]: E1126 06:40:31.573404 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1\": container with ID starting with 96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1 not found: ID does not exist" containerID="96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.573440 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1"} err="failed to get container status \"96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1\": rpc error: code = NotFound desc = could not find container \"96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1\": container with ID starting with 96bc82101e65020b61aae753d4a741d76648a8a2a0baddabf4afa436b79b21c1 not found: ID does not exist" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.573470 4775 scope.go:117] "RemoveContainer" containerID="f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87" Nov 26 06:40:31 crc kubenswrapper[4775]: E1126 06:40:31.573804 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87\": container with ID starting with f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87 not found: ID does not exist" containerID="f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.573832 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87"} err="failed to get container status \"f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87\": rpc error: code = NotFound desc = could not find container \"f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87\": container with ID starting with f55e524c432e6c2f3b45feeb9cc8236579159fced4cb4cff91f513cfd2d9ee87 not found: ID does not exist" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.573847 4775 scope.go:117] "RemoveContainer" containerID="42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637" Nov 26 06:40:31 crc kubenswrapper[4775]: E1126 06:40:31.574115 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637\": container with ID starting with 42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637 not found: ID does not exist" containerID="42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.574132 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637"} err="failed to get container status \"42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637\": rpc error: code = NotFound desc = could not find container \"42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637\": container with ID starting with 42dd15d3e824a28511744c9276db6d2a2c48995446a7b826815e298056d37637 not found: ID does not exist" Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.867241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:31 crc kubenswrapper[4775]: E1126 06:40:31.867414 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 06:40:31 crc kubenswrapper[4775]: E1126 06:40:31.867432 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 06:40:31 crc kubenswrapper[4775]: E1126 06:40:31.867486 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift podName:38df1b80-132c-467f-969c-d3798dd1f7d9 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:33.867472268 +0000 UTC m=+917.228776220 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift") pod "swift-storage-0" (UID: "38df1b80-132c-467f-969c-d3798dd1f7d9") : configmap "swift-ring-files" not found Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.878835 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2nsvj"] Nov 26 06:40:31 crc kubenswrapper[4775]: W1126 06:40:31.890024 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e5e0f59_7e39_43c9_b9d4_fe8dd69ec635.slice/crio-d1e5c1cd234801e5d973f978f1d6183a03143f208ce3df9bb31a5d2a3606bd8c WatchSource:0}: Error finding container d1e5c1cd234801e5d973f978f1d6183a03143f208ce3df9bb31a5d2a3606bd8c: Status 404 returned error can't find the container with id d1e5c1cd234801e5d973f978f1d6183a03143f208ce3df9bb31a5d2a3606bd8c Nov 26 06:40:31 crc kubenswrapper[4775]: I1126 06:40:31.905939 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldf4r"] Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.064290 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2nsvj" event={"ID":"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635","Type":"ContainerStarted","Data":"d1e5c1cd234801e5d973f978f1d6183a03143f208ce3df9bb31a5d2a3606bd8c"} Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.069236 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerStarted","Data":"976337157b3c33d5ba856ceebd84f2f538a087bf4b5e881bd8ef40251575bd0c"} Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.090548 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.117505 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=5.509174667 podStartE2EDuration="7.11748041s" podCreationTimestamp="2025-11-26 06:40:25 +0000 UTC" firstStartedPulling="2025-11-26 06:40:26.263576896 +0000 UTC m=+909.624880848" lastFinishedPulling="2025-11-26 06:40:27.871882639 +0000 UTC m=+911.233186591" observedRunningTime="2025-11-26 06:40:32.08964429 +0000 UTC m=+915.450948242" watchObservedRunningTime="2025-11-26 06:40:32.11748041 +0000 UTC m=+915.478784362" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.172628 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2f7p\" (UniqueName: \"kubernetes.io/projected/b2dd29bc-afa3-445a-be94-1b031b706b3d-kube-api-access-b2f7p\") pod \"b2dd29bc-afa3-445a-be94-1b031b706b3d\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.172673 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-nb\") pod \"b2dd29bc-afa3-445a-be94-1b031b706b3d\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.172731 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-dns-svc\") pod \"b2dd29bc-afa3-445a-be94-1b031b706b3d\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.172817 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-config\") pod \"b2dd29bc-afa3-445a-be94-1b031b706b3d\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.172843 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-sb\") pod \"b2dd29bc-afa3-445a-be94-1b031b706b3d\" (UID: \"b2dd29bc-afa3-445a-be94-1b031b706b3d\") " Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.176355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2dd29bc-afa3-445a-be94-1b031b706b3d-kube-api-access-b2f7p" (OuterVolumeSpecName: "kube-api-access-b2f7p") pod "b2dd29bc-afa3-445a-be94-1b031b706b3d" (UID: "b2dd29bc-afa3-445a-be94-1b031b706b3d"). InnerVolumeSpecName "kube-api-access-b2f7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.219345 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2dd29bc-afa3-445a-be94-1b031b706b3d" (UID: "b2dd29bc-afa3-445a-be94-1b031b706b3d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.219400 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2dd29bc-afa3-445a-be94-1b031b706b3d" (UID: "b2dd29bc-afa3-445a-be94-1b031b706b3d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.219412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2dd29bc-afa3-445a-be94-1b031b706b3d" (UID: "b2dd29bc-afa3-445a-be94-1b031b706b3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.222538 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-config" (OuterVolumeSpecName: "config") pod "b2dd29bc-afa3-445a-be94-1b031b706b3d" (UID: "b2dd29bc-afa3-445a-be94-1b031b706b3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.276121 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.276149 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.276160 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2f7p\" (UniqueName: \"kubernetes.io/projected/b2dd29bc-afa3-445a-be94-1b031b706b3d-kube-api-access-b2f7p\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.276169 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.276177 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2dd29bc-afa3-445a-be94-1b031b706b3d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.428330 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b814-account-create-update-pn25p"] Nov 26 06:40:32 crc kubenswrapper[4775]: E1126 06:40:32.428686 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerName="init" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.428697 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerName="init" Nov 26 06:40:32 crc kubenswrapper[4775]: E1126 06:40:32.428744 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerName="dnsmasq-dns" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.428750 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerName="dnsmasq-dns" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.428916 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" containerName="dnsmasq-dns" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.429480 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.455980 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.461979 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-w2kd5"] Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.463337 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.469202 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b814-account-create-update-pn25p"] Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.475928 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-w2kd5"] Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.478317 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e128aee-1376-4ee6-9bea-f5ff40e422d0-operator-scripts\") pod \"glance-b814-account-create-update-pn25p\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.478372 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n4rk\" (UniqueName: \"kubernetes.io/projected/6e128aee-1376-4ee6-9bea-f5ff40e422d0-kube-api-access-6n4rk\") pod \"glance-b814-account-create-update-pn25p\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.501058 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.579576 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2zt7\" (UniqueName: \"kubernetes.io/projected/aade9b70-d511-45e4-afcb-9887451cb4ed-kube-api-access-k2zt7\") pod \"glance-db-create-w2kd5\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.579624 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aade9b70-d511-45e4-afcb-9887451cb4ed-operator-scripts\") pod \"glance-db-create-w2kd5\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.579648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e128aee-1376-4ee6-9bea-f5ff40e422d0-operator-scripts\") pod \"glance-b814-account-create-update-pn25p\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.579676 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n4rk\" (UniqueName: \"kubernetes.io/projected/6e128aee-1376-4ee6-9bea-f5ff40e422d0-kube-api-access-6n4rk\") pod \"glance-b814-account-create-update-pn25p\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.580990 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e128aee-1376-4ee6-9bea-f5ff40e422d0-operator-scripts\") pod \"glance-b814-account-create-update-pn25p\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.598707 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n4rk\" (UniqueName: \"kubernetes.io/projected/6e128aee-1376-4ee6-9bea-f5ff40e422d0-kube-api-access-6n4rk\") pod \"glance-b814-account-create-update-pn25p\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.681184 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2zt7\" (UniqueName: \"kubernetes.io/projected/aade9b70-d511-45e4-afcb-9887451cb4ed-kube-api-access-k2zt7\") pod \"glance-db-create-w2kd5\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.681300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aade9b70-d511-45e4-afcb-9887451cb4ed-operator-scripts\") pod \"glance-db-create-w2kd5\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.682047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aade9b70-d511-45e4-afcb-9887451cb4ed-operator-scripts\") pod \"glance-db-create-w2kd5\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.707338 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2zt7\" (UniqueName: \"kubernetes.io/projected/aade9b70-d511-45e4-afcb-9887451cb4ed-kube-api-access-k2zt7\") pod \"glance-db-create-w2kd5\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.778900 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:32 crc kubenswrapper[4775]: I1126 06:40:32.793380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.082428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" event={"ID":"b2dd29bc-afa3-445a-be94-1b031b706b3d","Type":"ContainerDied","Data":"28f26e97931f9f820ad69dccc60c7881058156685c04bfa040d17c18807c83b2"} Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.082481 4775 scope.go:117] "RemoveContainer" containerID="0bfb6ef37bc33f00ae0701eda6eec33eb9de2c8db1ae6685bda1710d3536364e" Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.082512 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-mb28z" Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.084927 4775 generic.go:334] "Generic (PLEG): container finished" podID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerID="d3dd7e507ef69cea92870653a5ac20ba61f0f681bb9a2e2c3de82db6682d3a24" exitCode=0 Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.085018 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2nsvj" event={"ID":"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635","Type":"ContainerDied","Data":"d3dd7e507ef69cea92870653a5ac20ba61f0f681bb9a2e2c3de82db6682d3a24"} Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.087063 4775 generic.go:334] "Generic (PLEG): container finished" podID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerID="3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f" exitCode=0 Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.087093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerDied","Data":"3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f"} Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.105057 4775 scope.go:117] "RemoveContainer" containerID="9aca70db95c61976ec99d30423b7c9711fc91d7236748ac127bce2afc2fa69f5" Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.154273 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mb28z"] Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.161408 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-mb28z"] Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.267900 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b814-account-create-update-pn25p"] Nov 26 06:40:33 crc kubenswrapper[4775]: W1126 06:40:33.272245 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e128aee_1376_4ee6_9bea_f5ff40e422d0.slice/crio-0d98c1d3b87ea5f5b47e1145292cc20d1b0c7206fd61b7dc726ca4aa7e1bb53e WatchSource:0}: Error finding container 0d98c1d3b87ea5f5b47e1145292cc20d1b0c7206fd61b7dc726ca4aa7e1bb53e: Status 404 returned error can't find the container with id 0d98c1d3b87ea5f5b47e1145292cc20d1b0c7206fd61b7dc726ca4aa7e1bb53e Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.338950 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2dd29bc-afa3-445a-be94-1b031b706b3d" path="/var/lib/kubelet/pods/b2dd29bc-afa3-445a-be94-1b031b706b3d/volumes" Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.355561 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-w2kd5"] Nov 26 06:40:33 crc kubenswrapper[4775]: W1126 06:40:33.363121 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaade9b70_d511_45e4_afcb_9887451cb4ed.slice/crio-89022564ef5a64f55f833f074789db9f4a75910e7b7f1312f5346d1073e19f04 WatchSource:0}: Error finding container 89022564ef5a64f55f833f074789db9f4a75910e7b7f1312f5346d1073e19f04: Status 404 returned error can't find the container with id 89022564ef5a64f55f833f074789db9f4a75910e7b7f1312f5346d1073e19f04 Nov 26 06:40:33 crc kubenswrapper[4775]: I1126 06:40:33.908325 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:33 crc kubenswrapper[4775]: E1126 06:40:33.908534 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 06:40:33 crc kubenswrapper[4775]: E1126 06:40:33.908657 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 06:40:33 crc kubenswrapper[4775]: E1126 06:40:33.908708 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift podName:38df1b80-132c-467f-969c-d3798dd1f7d9 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:37.908694877 +0000 UTC m=+921.269998829 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift") pod "swift-storage-0" (UID: "38df1b80-132c-467f-969c-d3798dd1f7d9") : configmap "swift-ring-files" not found Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.089295 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-d8msj"] Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.090608 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.095136 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.095136 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.095386 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.097801 4775 generic.go:334] "Generic (PLEG): container finished" podID="aade9b70-d511-45e4-afcb-9887451cb4ed" containerID="6c51504d44cfa68b422120558f2b17abc586f5fabec98b08f856dc57e5118dab" exitCode=0 Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.097903 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-w2kd5" event={"ID":"aade9b70-d511-45e4-afcb-9887451cb4ed","Type":"ContainerDied","Data":"6c51504d44cfa68b422120558f2b17abc586f5fabec98b08f856dc57e5118dab"} Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.097929 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-w2kd5" event={"ID":"aade9b70-d511-45e4-afcb-9887451cb4ed","Type":"ContainerStarted","Data":"89022564ef5a64f55f833f074789db9f4a75910e7b7f1312f5346d1073e19f04"} Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.099747 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2nsvj" event={"ID":"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635","Type":"ContainerStarted","Data":"2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296"} Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.099875 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.101557 4775 generic.go:334] "Generic (PLEG): container finished" podID="6e128aee-1376-4ee6-9bea-f5ff40e422d0" containerID="4383c34721635f3ef8676856d6f0165c4adeb93596757a9dd69c46556d6ad68c" exitCode=0 Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.101591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b814-account-create-update-pn25p" event={"ID":"6e128aee-1376-4ee6-9bea-f5ff40e422d0","Type":"ContainerDied","Data":"4383c34721635f3ef8676856d6f0165c4adeb93596757a9dd69c46556d6ad68c"} Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.101612 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b814-account-create-update-pn25p" event={"ID":"6e128aee-1376-4ee6-9bea-f5ff40e422d0","Type":"ContainerStarted","Data":"0d98c1d3b87ea5f5b47e1145292cc20d1b0c7206fd61b7dc726ca4aa7e1bb53e"} Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.104567 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d8msj"] Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.185541 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-2nsvj" podStartSLOduration=6.185524381 podStartE2EDuration="6.185524381s" podCreationTimestamp="2025-11-26 06:40:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:40:34.180174877 +0000 UTC m=+917.541478829" watchObservedRunningTime="2025-11-26 06:40:34.185524381 +0000 UTC m=+917.546828333" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.212813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-combined-ca-bundle\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.213008 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-ring-data-devices\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.213195 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-scripts\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.213225 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-dispersionconf\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.213344 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq6g5\" (UniqueName: \"kubernetes.io/projected/fd3c60bd-b664-43b1-b669-3ef640959006-kube-api-access-jq6g5\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.213437 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-swiftconf\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.213461 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fd3c60bd-b664-43b1-b669-3ef640959006-etc-swift\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315490 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-ring-data-devices\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-scripts\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315663 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-dispersionconf\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315764 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq6g5\" (UniqueName: \"kubernetes.io/projected/fd3c60bd-b664-43b1-b669-3ef640959006-kube-api-access-jq6g5\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-swiftconf\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315886 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fd3c60bd-b664-43b1-b669-3ef640959006-etc-swift\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.315957 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-combined-ca-bundle\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.316472 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-scripts\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.316484 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fd3c60bd-b664-43b1-b669-3ef640959006-etc-swift\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.316512 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-ring-data-devices\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.321546 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-dispersionconf\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.321969 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-swiftconf\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.326807 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-combined-ca-bundle\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.337509 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq6g5\" (UniqueName: \"kubernetes.io/projected/fd3c60bd-b664-43b1-b669-3ef640959006-kube-api-access-jq6g5\") pod \"swift-ring-rebalance-d8msj\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.407396 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:34 crc kubenswrapper[4775]: I1126 06:40:34.693775 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-d8msj"] Nov 26 06:40:34 crc kubenswrapper[4775]: W1126 06:40:34.711246 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd3c60bd_b664_43b1_b669_3ef640959006.slice/crio-2808165fe0dd2c32947b84019c507af32b1a337d587ed7291c288568222788ed WatchSource:0}: Error finding container 2808165fe0dd2c32947b84019c507af32b1a337d587ed7291c288568222788ed: Status 404 returned error can't find the container with id 2808165fe0dd2c32947b84019c507af32b1a337d587ed7291c288568222788ed Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.019992 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hb7d2"] Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.020296 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hb7d2" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="registry-server" containerID="cri-o://1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5" gracePeriod=2 Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.110569 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d8msj" event={"ID":"fd3c60bd-b664-43b1-b669-3ef640959006","Type":"ContainerStarted","Data":"2808165fe0dd2c32947b84019c507af32b1a337d587ed7291c288568222788ed"} Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.114488 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerStarted","Data":"747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036"} Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.645732 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.652162 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.655371 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749751 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2zt7\" (UniqueName: \"kubernetes.io/projected/aade9b70-d511-45e4-afcb-9887451cb4ed-kube-api-access-k2zt7\") pod \"aade9b70-d511-45e4-afcb-9887451cb4ed\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749804 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aade9b70-d511-45e4-afcb-9887451cb4ed-operator-scripts\") pod \"aade9b70-d511-45e4-afcb-9887451cb4ed\" (UID: \"aade9b70-d511-45e4-afcb-9887451cb4ed\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749858 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f7rn\" (UniqueName: \"kubernetes.io/projected/46ce0a21-d5df-44fb-96cc-8e008b5e989a-kube-api-access-5f7rn\") pod \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749876 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e128aee-1376-4ee6-9bea-f5ff40e422d0-operator-scripts\") pod \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749912 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-utilities\") pod \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n4rk\" (UniqueName: \"kubernetes.io/projected/6e128aee-1376-4ee6-9bea-f5ff40e422d0-kube-api-access-6n4rk\") pod \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\" (UID: \"6e128aee-1376-4ee6-9bea-f5ff40e422d0\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.749993 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-catalog-content\") pod \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\" (UID: \"46ce0a21-d5df-44fb-96cc-8e008b5e989a\") " Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.750943 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e128aee-1376-4ee6-9bea-f5ff40e422d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6e128aee-1376-4ee6-9bea-f5ff40e422d0" (UID: "6e128aee-1376-4ee6-9bea-f5ff40e422d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.751321 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aade9b70-d511-45e4-afcb-9887451cb4ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aade9b70-d511-45e4-afcb-9887451cb4ed" (UID: "aade9b70-d511-45e4-afcb-9887451cb4ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.756254 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aade9b70-d511-45e4-afcb-9887451cb4ed-kube-api-access-k2zt7" (OuterVolumeSpecName: "kube-api-access-k2zt7") pod "aade9b70-d511-45e4-afcb-9887451cb4ed" (UID: "aade9b70-d511-45e4-afcb-9887451cb4ed"). InnerVolumeSpecName "kube-api-access-k2zt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.757222 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-utilities" (OuterVolumeSpecName: "utilities") pod "46ce0a21-d5df-44fb-96cc-8e008b5e989a" (UID: "46ce0a21-d5df-44fb-96cc-8e008b5e989a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.760213 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e128aee-1376-4ee6-9bea-f5ff40e422d0-kube-api-access-6n4rk" (OuterVolumeSpecName: "kube-api-access-6n4rk") pod "6e128aee-1376-4ee6-9bea-f5ff40e422d0" (UID: "6e128aee-1376-4ee6-9bea-f5ff40e422d0"). InnerVolumeSpecName "kube-api-access-6n4rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.761088 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46ce0a21-d5df-44fb-96cc-8e008b5e989a-kube-api-access-5f7rn" (OuterVolumeSpecName: "kube-api-access-5f7rn") pod "46ce0a21-d5df-44fb-96cc-8e008b5e989a" (UID: "46ce0a21-d5df-44fb-96cc-8e008b5e989a"). InnerVolumeSpecName "kube-api-access-5f7rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.813338 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.815068 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46ce0a21-d5df-44fb-96cc-8e008b5e989a" (UID: "46ce0a21-d5df-44fb-96cc-8e008b5e989a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860070 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860116 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n4rk\" (UniqueName: \"kubernetes.io/projected/6e128aee-1376-4ee6-9bea-f5ff40e422d0-kube-api-access-6n4rk\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860150 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46ce0a21-d5df-44fb-96cc-8e008b5e989a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860162 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2zt7\" (UniqueName: \"kubernetes.io/projected/aade9b70-d511-45e4-afcb-9887451cb4ed-kube-api-access-k2zt7\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860175 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aade9b70-d511-45e4-afcb-9887451cb4ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860186 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f7rn\" (UniqueName: \"kubernetes.io/projected/46ce0a21-d5df-44fb-96cc-8e008b5e989a-kube-api-access-5f7rn\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:35 crc kubenswrapper[4775]: I1126 06:40:35.860198 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6e128aee-1376-4ee6-9bea-f5ff40e422d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.123671 4775 generic.go:334] "Generic (PLEG): container finished" podID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerID="1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5" exitCode=0 Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.124926 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hb7d2" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.124911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerDied","Data":"1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5"} Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.125733 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hb7d2" event={"ID":"46ce0a21-d5df-44fb-96cc-8e008b5e989a","Type":"ContainerDied","Data":"c09419cada9c02b8ea8829f97df92f1969477bc59ddc4a4787838ab1e2bbd722"} Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.125778 4775 scope.go:117] "RemoveContainer" containerID="1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.129255 4775 generic.go:334] "Generic (PLEG): container finished" podID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerID="747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036" exitCode=0 Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.129350 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerDied","Data":"747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036"} Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.130936 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b814-account-create-update-pn25p" event={"ID":"6e128aee-1376-4ee6-9bea-f5ff40e422d0","Type":"ContainerDied","Data":"0d98c1d3b87ea5f5b47e1145292cc20d1b0c7206fd61b7dc726ca4aa7e1bb53e"} Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.130973 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d98c1d3b87ea5f5b47e1145292cc20d1b0c7206fd61b7dc726ca4aa7e1bb53e" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.130948 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b814-account-create-update-pn25p" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.132842 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-w2kd5" event={"ID":"aade9b70-d511-45e4-afcb-9887451cb4ed","Type":"ContainerDied","Data":"89022564ef5a64f55f833f074789db9f4a75910e7b7f1312f5346d1073e19f04"} Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.132876 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89022564ef5a64f55f833f074789db9f4a75910e7b7f1312f5346d1073e19f04" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.132927 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-w2kd5" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.182877 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hb7d2"] Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.189462 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hb7d2"] Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.767312 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g6rtl"] Nov 26 06:40:36 crc kubenswrapper[4775]: E1126 06:40:36.767693 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e128aee-1376-4ee6-9bea-f5ff40e422d0" containerName="mariadb-account-create-update" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.767736 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e128aee-1376-4ee6-9bea-f5ff40e422d0" containerName="mariadb-account-create-update" Nov 26 06:40:36 crc kubenswrapper[4775]: E1126 06:40:36.767756 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="extract-utilities" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.767762 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="extract-utilities" Nov 26 06:40:36 crc kubenswrapper[4775]: E1126 06:40:36.767778 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="registry-server" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.767784 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="registry-server" Nov 26 06:40:36 crc kubenswrapper[4775]: E1126 06:40:36.767794 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aade9b70-d511-45e4-afcb-9887451cb4ed" containerName="mariadb-database-create" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.767799 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aade9b70-d511-45e4-afcb-9887451cb4ed" containerName="mariadb-database-create" Nov 26 06:40:36 crc kubenswrapper[4775]: E1126 06:40:36.767810 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="extract-content" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.767834 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="extract-content" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.768058 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" containerName="registry-server" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.768071 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e128aee-1376-4ee6-9bea-f5ff40e422d0" containerName="mariadb-account-create-update" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.768083 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="aade9b70-d511-45e4-afcb-9887451cb4ed" containerName="mariadb-database-create" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.768597 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.779977 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g6rtl"] Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.865316 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b988-account-create-update-7tg4l"] Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.866306 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.868431 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.873035 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b988-account-create-update-7tg4l"] Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.882993 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-operator-scripts\") pod \"keystone-db-create-g6rtl\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.883152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77nnx\" (UniqueName: \"kubernetes.io/projected/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-kube-api-access-77nnx\") pod \"keystone-db-create-g6rtl\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.984518 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tqpm\" (UniqueName: \"kubernetes.io/projected/73081230-d192-4993-9434-9c4c6a5533f9-kube-api-access-2tqpm\") pod \"keystone-b988-account-create-update-7tg4l\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.984566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-operator-scripts\") pod \"keystone-db-create-g6rtl\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.984653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77nnx\" (UniqueName: \"kubernetes.io/projected/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-kube-api-access-77nnx\") pod \"keystone-db-create-g6rtl\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.984691 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73081230-d192-4993-9434-9c4c6a5533f9-operator-scripts\") pod \"keystone-b988-account-create-update-7tg4l\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:36 crc kubenswrapper[4775]: I1126 06:40:36.985739 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-operator-scripts\") pod \"keystone-db-create-g6rtl\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.007003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77nnx\" (UniqueName: \"kubernetes.io/projected/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-kube-api-access-77nnx\") pod \"keystone-db-create-g6rtl\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.085629 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tqpm\" (UniqueName: \"kubernetes.io/projected/73081230-d192-4993-9434-9c4c6a5533f9-kube-api-access-2tqpm\") pod \"keystone-b988-account-create-update-7tg4l\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.085779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73081230-d192-4993-9434-9c4c6a5533f9-operator-scripts\") pod \"keystone-b988-account-create-update-7tg4l\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.086584 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73081230-d192-4993-9434-9c4c6a5533f9-operator-scripts\") pod \"keystone-b988-account-create-update-7tg4l\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.090277 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.115257 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tqpm\" (UniqueName: \"kubernetes.io/projected/73081230-d192-4993-9434-9c4c6a5533f9-kube-api-access-2tqpm\") pod \"keystone-b988-account-create-update-7tg4l\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.169549 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5sdcc"] Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.171486 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.177173 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-097e-account-create-update-khn5j"] Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.178774 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.181185 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.183155 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.196968 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5sdcc"] Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.203244 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-097e-account-create-update-khn5j"] Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.289360 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k2h9\" (UniqueName: \"kubernetes.io/projected/c54287b0-da13-43bc-9a94-d5fa303b885e-kube-api-access-5k2h9\") pod \"placement-db-create-5sdcc\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.289417 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khhjk\" (UniqueName: \"kubernetes.io/projected/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-kube-api-access-khhjk\") pod \"placement-097e-account-create-update-khn5j\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.289560 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-operator-scripts\") pod \"placement-097e-account-create-update-khn5j\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.289632 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54287b0-da13-43bc-9a94-d5fa303b885e-operator-scripts\") pod \"placement-db-create-5sdcc\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.344772 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46ce0a21-d5df-44fb-96cc-8e008b5e989a" path="/var/lib/kubelet/pods/46ce0a21-d5df-44fb-96cc-8e008b5e989a/volumes" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.391349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-operator-scripts\") pod \"placement-097e-account-create-update-khn5j\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.391850 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54287b0-da13-43bc-9a94-d5fa303b885e-operator-scripts\") pod \"placement-db-create-5sdcc\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.391906 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k2h9\" (UniqueName: \"kubernetes.io/projected/c54287b0-da13-43bc-9a94-d5fa303b885e-kube-api-access-5k2h9\") pod \"placement-db-create-5sdcc\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.391922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khhjk\" (UniqueName: \"kubernetes.io/projected/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-kube-api-access-khhjk\") pod \"placement-097e-account-create-update-khn5j\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.393698 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-operator-scripts\") pod \"placement-097e-account-create-update-khn5j\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.394488 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54287b0-da13-43bc-9a94-d5fa303b885e-operator-scripts\") pod \"placement-db-create-5sdcc\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.413909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k2h9\" (UniqueName: \"kubernetes.io/projected/c54287b0-da13-43bc-9a94-d5fa303b885e-kube-api-access-5k2h9\") pod \"placement-db-create-5sdcc\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.415852 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khhjk\" (UniqueName: \"kubernetes.io/projected/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-kube-api-access-khhjk\") pod \"placement-097e-account-create-update-khn5j\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.508487 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.513985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.667801 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-bnfj2"] Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.669008 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.671794 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qtdst" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.671835 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.703258 4775 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podf6381ad9-d93b-4f52-a5ae-4d694f4db96d"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podf6381ad9-d93b-4f52-a5ae-4d694f4db96d] : Timed out while waiting for systemd to remove kubepods-besteffort-podf6381ad9_d93b_4f52_a5ae_4d694f4db96d.slice" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.719146 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bnfj2"] Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.798657 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-combined-ca-bundle\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.798687 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldkxb\" (UniqueName: \"kubernetes.io/projected/06776a0e-14fb-411c-b70d-861de9969bcd-kube-api-access-ldkxb\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.798776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-db-sync-config-data\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.798878 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-config-data\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.900235 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-combined-ca-bundle\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.901658 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldkxb\" (UniqueName: \"kubernetes.io/projected/06776a0e-14fb-411c-b70d-861de9969bcd-kube-api-access-ldkxb\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.901807 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-db-sync-config-data\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.901974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-config-data\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.904694 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-combined-ca-bundle\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.905655 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-db-sync-config-data\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.908125 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-config-data\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.920580 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldkxb\" (UniqueName: \"kubernetes.io/projected/06776a0e-14fb-411c-b70d-861de9969bcd-kube-api-access-ldkxb\") pod \"glance-db-sync-bnfj2\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:37 crc kubenswrapper[4775]: I1126 06:40:37.990164 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bnfj2" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.004572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:38 crc kubenswrapper[4775]: E1126 06:40:38.005123 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 06:40:38 crc kubenswrapper[4775]: E1126 06:40:38.005153 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 06:40:38 crc kubenswrapper[4775]: E1126 06:40:38.005218 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift podName:38df1b80-132c-467f-969c-d3798dd1f7d9 nodeName:}" failed. No retries permitted until 2025-11-26 06:40:46.005195794 +0000 UTC m=+929.366499746 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift") pod "swift-storage-0" (UID: "38df1b80-132c-467f-969c-d3798dd1f7d9") : configmap "swift-ring-files" not found Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.358240 4775 scope.go:117] "RemoveContainer" containerID="00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.557956 4775 scope.go:117] "RemoveContainer" containerID="f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.599004 4775 scope.go:117] "RemoveContainer" containerID="1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5" Nov 26 06:40:38 crc kubenswrapper[4775]: E1126 06:40:38.599506 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5\": container with ID starting with 1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5 not found: ID does not exist" containerID="1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.599536 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5"} err="failed to get container status \"1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5\": rpc error: code = NotFound desc = could not find container \"1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5\": container with ID starting with 1ea4dadece08e8392ecf7be4eb82bf20c93a1e2be7402858eaffa093e55c6cb5 not found: ID does not exist" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.599570 4775 scope.go:117] "RemoveContainer" containerID="00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a" Nov 26 06:40:38 crc kubenswrapper[4775]: E1126 06:40:38.599882 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a\": container with ID starting with 00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a not found: ID does not exist" containerID="00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.599919 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a"} err="failed to get container status \"00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a\": rpc error: code = NotFound desc = could not find container \"00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a\": container with ID starting with 00ea5ea399f6d60a0e3907644070af811311014acc0bf3359924f809688d4e5a not found: ID does not exist" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.599933 4775 scope.go:117] "RemoveContainer" containerID="f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2" Nov 26 06:40:38 crc kubenswrapper[4775]: E1126 06:40:38.600158 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2\": container with ID starting with f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2 not found: ID does not exist" containerID="f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.600175 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2"} err="failed to get container status \"f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2\": rpc error: code = NotFound desc = could not find container \"f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2\": container with ID starting with f7d91e6cf63540852b54e6dc83dd72316c913b0e00e40448d8e05a250bdb84b2 not found: ID does not exist" Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.855575 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5sdcc"] Nov 26 06:40:38 crc kubenswrapper[4775]: I1126 06:40:38.923034 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g6rtl"] Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.010029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-097e-account-create-update-khn5j"] Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.027385 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b988-account-create-update-7tg4l"] Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.037379 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bnfj2"] Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.170413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerStarted","Data":"d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.176233 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g6rtl" event={"ID":"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2","Type":"ContainerStarted","Data":"985c19991e9b6dd4c534806141b3bcb960c1a76c13d013237b6faebc845f8376"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.179933 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b988-account-create-update-7tg4l" event={"ID":"73081230-d192-4993-9434-9c4c6a5533f9","Type":"ContainerStarted","Data":"893798eb94929b6200ddbd4bc3b0eba40bf6160321c1a3b899f295c5bd9660eb"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.181763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d8msj" event={"ID":"fd3c60bd-b664-43b1-b669-3ef640959006","Type":"ContainerStarted","Data":"cc0d270194d0d930bb46b0d6f9ff37aa231ef84158d3e996e3ef4b99430a135e"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.185973 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bnfj2" event={"ID":"06776a0e-14fb-411c-b70d-861de9969bcd","Type":"ContainerStarted","Data":"4290294b61fed7ea376a6005f60be7a8f5f314b58028aabf6a9cc63f2d377655"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.191126 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ldf4r" podStartSLOduration=3.8427201220000002 podStartE2EDuration="9.191105984s" podCreationTimestamp="2025-11-26 06:40:30 +0000 UTC" firstStartedPulling="2025-11-26 06:40:33.088585747 +0000 UTC m=+916.449889699" lastFinishedPulling="2025-11-26 06:40:38.436971599 +0000 UTC m=+921.798275561" observedRunningTime="2025-11-26 06:40:39.187115517 +0000 UTC m=+922.548419459" watchObservedRunningTime="2025-11-26 06:40:39.191105984 +0000 UTC m=+922.552409946" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.192508 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5sdcc" event={"ID":"c54287b0-da13-43bc-9a94-d5fa303b885e","Type":"ContainerStarted","Data":"53d27861306ed497ddfe36a93ebdd902141ad3b9522aa0335274ad3b79657c7c"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.192547 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5sdcc" event={"ID":"c54287b0-da13-43bc-9a94-d5fa303b885e","Type":"ContainerStarted","Data":"c0f8bd1b0a631b1fe1cf34148ca479cf1b5db3c0fd1202e4f7005a157e8024e1"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.201188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097e-account-create-update-khn5j" event={"ID":"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab","Type":"ContainerStarted","Data":"1d3a9caa918fa68c321d47bfd5cac11b4e766c717fdbb502af47ebe74551f769"} Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.203596 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-d8msj" podStartSLOduration=1.480265591 podStartE2EDuration="5.20358322s" podCreationTimestamp="2025-11-26 06:40:34 +0000 UTC" firstStartedPulling="2025-11-26 06:40:34.714301498 +0000 UTC m=+918.075605450" lastFinishedPulling="2025-11-26 06:40:38.437619127 +0000 UTC m=+921.798923079" observedRunningTime="2025-11-26 06:40:39.202804209 +0000 UTC m=+922.564108161" watchObservedRunningTime="2025-11-26 06:40:39.20358322 +0000 UTC m=+922.564887172" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.223448 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-5sdcc" podStartSLOduration=2.223429465 podStartE2EDuration="2.223429465s" podCreationTimestamp="2025-11-26 06:40:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:40:39.213300282 +0000 UTC m=+922.574604234" watchObservedRunningTime="2025-11-26 06:40:39.223429465 +0000 UTC m=+922.584733417" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.391581 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.478890 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nfsh5"] Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.479157 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" podUID="01958f53-a2a9-4783-9669-abd002d57a66" containerName="dnsmasq-dns" containerID="cri-o://a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c" gracePeriod=10 Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.897966 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.941512 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-dns-svc\") pod \"01958f53-a2a9-4783-9669-abd002d57a66\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.941630 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br7q6\" (UniqueName: \"kubernetes.io/projected/01958f53-a2a9-4783-9669-abd002d57a66-kube-api-access-br7q6\") pod \"01958f53-a2a9-4783-9669-abd002d57a66\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.941693 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-config\") pod \"01958f53-a2a9-4783-9669-abd002d57a66\" (UID: \"01958f53-a2a9-4783-9669-abd002d57a66\") " Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.946784 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01958f53-a2a9-4783-9669-abd002d57a66-kube-api-access-br7q6" (OuterVolumeSpecName: "kube-api-access-br7q6") pod "01958f53-a2a9-4783-9669-abd002d57a66" (UID: "01958f53-a2a9-4783-9669-abd002d57a66"). InnerVolumeSpecName "kube-api-access-br7q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.988436 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01958f53-a2a9-4783-9669-abd002d57a66" (UID: "01958f53-a2a9-4783-9669-abd002d57a66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:39 crc kubenswrapper[4775]: I1126 06:40:39.988925 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-config" (OuterVolumeSpecName: "config") pod "01958f53-a2a9-4783-9669-abd002d57a66" (UID: "01958f53-a2a9-4783-9669-abd002d57a66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.044493 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.044544 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br7q6\" (UniqueName: \"kubernetes.io/projected/01958f53-a2a9-4783-9669-abd002d57a66-kube-api-access-br7q6\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.044560 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01958f53-a2a9-4783-9669-abd002d57a66-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.219649 4775 generic.go:334] "Generic (PLEG): container finished" podID="e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" containerID="163fefffded403580198d0723221e6aef8818363230465cdd4a6e640b2e9ef3a" exitCode=0 Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.219759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097e-account-create-update-khn5j" event={"ID":"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab","Type":"ContainerDied","Data":"163fefffded403580198d0723221e6aef8818363230465cdd4a6e640b2e9ef3a"} Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.221595 4775 generic.go:334] "Generic (PLEG): container finished" podID="b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" containerID="0e37d7d74b50450e83ffb9050af536a4e36a5438d7290eb9958067ab87c28913" exitCode=0 Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.221651 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g6rtl" event={"ID":"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2","Type":"ContainerDied","Data":"0e37d7d74b50450e83ffb9050af536a4e36a5438d7290eb9958067ab87c28913"} Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.223489 4775 generic.go:334] "Generic (PLEG): container finished" podID="01958f53-a2a9-4783-9669-abd002d57a66" containerID="a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c" exitCode=0 Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.223549 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.223557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" event={"ID":"01958f53-a2a9-4783-9669-abd002d57a66","Type":"ContainerDied","Data":"a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c"} Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.223680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-nfsh5" event={"ID":"01958f53-a2a9-4783-9669-abd002d57a66","Type":"ContainerDied","Data":"af8796a852a66cc2dd686cabb4b075976b7dd9162567689ebd90561e658424dc"} Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.223705 4775 scope.go:117] "RemoveContainer" containerID="a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.225193 4775 generic.go:334] "Generic (PLEG): container finished" podID="73081230-d192-4993-9434-9c4c6a5533f9" containerID="92412a8829994c6bfbd9ee61f3fc8c92186e5345eb38bfadade4767032bc9d14" exitCode=0 Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.225250 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b988-account-create-update-7tg4l" event={"ID":"73081230-d192-4993-9434-9c4c6a5533f9","Type":"ContainerDied","Data":"92412a8829994c6bfbd9ee61f3fc8c92186e5345eb38bfadade4767032bc9d14"} Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.226456 4775 generic.go:334] "Generic (PLEG): container finished" podID="c54287b0-da13-43bc-9a94-d5fa303b885e" containerID="53d27861306ed497ddfe36a93ebdd902141ad3b9522aa0335274ad3b79657c7c" exitCode=0 Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.227410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5sdcc" event={"ID":"c54287b0-da13-43bc-9a94-d5fa303b885e","Type":"ContainerDied","Data":"53d27861306ed497ddfe36a93ebdd902141ad3b9522aa0335274ad3b79657c7c"} Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.250329 4775 scope.go:117] "RemoveContainer" containerID="39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.275746 4775 scope.go:117] "RemoveContainer" containerID="a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c" Nov 26 06:40:40 crc kubenswrapper[4775]: E1126 06:40:40.276250 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c\": container with ID starting with a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c not found: ID does not exist" containerID="a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.276380 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c"} err="failed to get container status \"a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c\": rpc error: code = NotFound desc = could not find container \"a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c\": container with ID starting with a2c691dc21c5baaf1f7e1c604486a309ca3d38a3754589a8f6c2db72ab08e45c not found: ID does not exist" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.276555 4775 scope.go:117] "RemoveContainer" containerID="39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85" Nov 26 06:40:40 crc kubenswrapper[4775]: E1126 06:40:40.283858 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85\": container with ID starting with 39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85 not found: ID does not exist" containerID="39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.283890 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85"} err="failed to get container status \"39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85\": rpc error: code = NotFound desc = could not find container \"39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85\": container with ID starting with 39c659aaaa5bc13075be97fb1990ade89aab8b3e9370efff6c160b0c366d6c85 not found: ID does not exist" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.301797 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nfsh5"] Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.308495 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-nfsh5"] Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.807899 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.808261 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.863749 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:40 crc kubenswrapper[4775]: I1126 06:40:40.873217 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.370013 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01958f53-a2a9-4783-9669-abd002d57a66" path="/var/lib/kubelet/pods/01958f53-a2a9-4783-9669-abd002d57a66/volumes" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.637583 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.671449 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k2h9\" (UniqueName: \"kubernetes.io/projected/c54287b0-da13-43bc-9a94-d5fa303b885e-kube-api-access-5k2h9\") pod \"c54287b0-da13-43bc-9a94-d5fa303b885e\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.671519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54287b0-da13-43bc-9a94-d5fa303b885e-operator-scripts\") pod \"c54287b0-da13-43bc-9a94-d5fa303b885e\" (UID: \"c54287b0-da13-43bc-9a94-d5fa303b885e\") " Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.672604 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c54287b0-da13-43bc-9a94-d5fa303b885e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c54287b0-da13-43bc-9a94-d5fa303b885e" (UID: "c54287b0-da13-43bc-9a94-d5fa303b885e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.678974 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c54287b0-da13-43bc-9a94-d5fa303b885e-kube-api-access-5k2h9" (OuterVolumeSpecName: "kube-api-access-5k2h9") pod "c54287b0-da13-43bc-9a94-d5fa303b885e" (UID: "c54287b0-da13-43bc-9a94-d5fa303b885e"). InnerVolumeSpecName "kube-api-access-5k2h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.775784 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k2h9\" (UniqueName: \"kubernetes.io/projected/c54287b0-da13-43bc-9a94-d5fa303b885e-kube-api-access-5k2h9\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.775824 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c54287b0-da13-43bc-9a94-d5fa303b885e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.885909 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.897231 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:41 crc kubenswrapper[4775]: I1126 06:40:41.913573 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.094443 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-operator-scripts\") pod \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.094520 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73081230-d192-4993-9434-9c4c6a5533f9-operator-scripts\") pod \"73081230-d192-4993-9434-9c4c6a5533f9\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.094800 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khhjk\" (UniqueName: \"kubernetes.io/projected/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-kube-api-access-khhjk\") pod \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.094895 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-operator-scripts\") pod \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\" (UID: \"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab\") " Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.094947 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77nnx\" (UniqueName: \"kubernetes.io/projected/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-kube-api-access-77nnx\") pod \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\" (UID: \"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2\") " Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.094974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tqpm\" (UniqueName: \"kubernetes.io/projected/73081230-d192-4993-9434-9c4c6a5533f9-kube-api-access-2tqpm\") pod \"73081230-d192-4993-9434-9c4c6a5533f9\" (UID: \"73081230-d192-4993-9434-9c4c6a5533f9\") " Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.096527 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" (UID: "b8957b6f-3c9b-4562-9720-c9dbc09d3ee2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.096601 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" (UID: "e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.096840 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73081230-d192-4993-9434-9c4c6a5533f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73081230-d192-4993-9434-9c4c6a5533f9" (UID: "73081230-d192-4993-9434-9c4c6a5533f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.099488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-kube-api-access-khhjk" (OuterVolumeSpecName: "kube-api-access-khhjk") pod "e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" (UID: "e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab"). InnerVolumeSpecName "kube-api-access-khhjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.099921 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73081230-d192-4993-9434-9c4c6a5533f9-kube-api-access-2tqpm" (OuterVolumeSpecName: "kube-api-access-2tqpm") pod "73081230-d192-4993-9434-9c4c6a5533f9" (UID: "73081230-d192-4993-9434-9c4c6a5533f9"). InnerVolumeSpecName "kube-api-access-2tqpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.100918 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-kube-api-access-77nnx" (OuterVolumeSpecName: "kube-api-access-77nnx") pod "b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" (UID: "b8957b6f-3c9b-4562-9720-c9dbc09d3ee2"). InnerVolumeSpecName "kube-api-access-77nnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.197981 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khhjk\" (UniqueName: \"kubernetes.io/projected/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-kube-api-access-khhjk\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.198024 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.198054 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77nnx\" (UniqueName: \"kubernetes.io/projected/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-kube-api-access-77nnx\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.198065 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tqpm\" (UniqueName: \"kubernetes.io/projected/73081230-d192-4993-9434-9c4c6a5533f9-kube-api-access-2tqpm\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.198077 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.198086 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73081230-d192-4993-9434-9c4c6a5533f9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.251742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5sdcc" event={"ID":"c54287b0-da13-43bc-9a94-d5fa303b885e","Type":"ContainerDied","Data":"c0f8bd1b0a631b1fe1cf34148ca479cf1b5db3c0fd1202e4f7005a157e8024e1"} Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.251782 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f8bd1b0a631b1fe1cf34148ca479cf1b5db3c0fd1202e4f7005a157e8024e1" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.251838 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5sdcc" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.253241 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-097e-account-create-update-khn5j" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.253263 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-097e-account-create-update-khn5j" event={"ID":"e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab","Type":"ContainerDied","Data":"1d3a9caa918fa68c321d47bfd5cac11b4e766c717fdbb502af47ebe74551f769"} Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.253327 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d3a9caa918fa68c321d47bfd5cac11b4e766c717fdbb502af47ebe74551f769" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.254836 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g6rtl" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.255000 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g6rtl" event={"ID":"b8957b6f-3c9b-4562-9720-c9dbc09d3ee2","Type":"ContainerDied","Data":"985c19991e9b6dd4c534806141b3bcb960c1a76c13d013237b6faebc845f8376"} Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.255028 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="985c19991e9b6dd4c534806141b3bcb960c1a76c13d013237b6faebc845f8376" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.256988 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b988-account-create-update-7tg4l" Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.261982 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b988-account-create-update-7tg4l" event={"ID":"73081230-d192-4993-9434-9c4c6a5533f9","Type":"ContainerDied","Data":"893798eb94929b6200ddbd4bc3b0eba40bf6160321c1a3b899f295c5bd9660eb"} Nov 26 06:40:42 crc kubenswrapper[4775]: I1126 06:40:42.262024 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="893798eb94929b6200ddbd4bc3b0eba40bf6160321c1a3b899f295c5bd9660eb" Nov 26 06:40:46 crc kubenswrapper[4775]: I1126 06:40:46.076659 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:40:46 crc kubenswrapper[4775]: E1126 06:40:46.076916 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 26 06:40:46 crc kubenswrapper[4775]: E1126 06:40:46.077133 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 26 06:40:46 crc kubenswrapper[4775]: E1126 06:40:46.077186 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift podName:38df1b80-132c-467f-969c-d3798dd1f7d9 nodeName:}" failed. No retries permitted until 2025-11-26 06:41:02.077170339 +0000 UTC m=+945.438474291 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift") pod "swift-storage-0" (UID: "38df1b80-132c-467f-969c-d3798dd1f7d9") : configmap "swift-ring-files" not found Nov 26 06:40:47 crc kubenswrapper[4775]: I1126 06:40:47.299284 4775 generic.go:334] "Generic (PLEG): container finished" podID="fd3c60bd-b664-43b1-b669-3ef640959006" containerID="cc0d270194d0d930bb46b0d6f9ff37aa231ef84158d3e996e3ef4b99430a135e" exitCode=0 Nov 26 06:40:47 crc kubenswrapper[4775]: I1126 06:40:47.299324 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d8msj" event={"ID":"fd3c60bd-b664-43b1-b669-3ef640959006","Type":"ContainerDied","Data":"cc0d270194d0d930bb46b0d6f9ff37aa231ef84158d3e996e3ef4b99430a135e"} Nov 26 06:40:47 crc kubenswrapper[4775]: I1126 06:40:47.740319 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k4mqj" podUID="d30382f9-28bd-4ab7-8fec-1f8932b3f804" containerName="ovn-controller" probeResult="failure" output=< Nov 26 06:40:47 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 26 06:40:47 crc kubenswrapper[4775]: > Nov 26 06:40:50 crc kubenswrapper[4775]: I1126 06:40:50.864734 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:50 crc kubenswrapper[4775]: I1126 06:40:50.912570 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldf4r"] Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.280586 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.337380 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-d8msj" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.344266 4775 generic.go:334] "Generic (PLEG): container finished" podID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerID="bcdd2a68e5172c990298b42de9bed69bd8c00801bcd6c8e980fd5d7cc6fafff6" exitCode=0 Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.346491 4775 generic.go:334] "Generic (PLEG): container finished" podID="63f863dc-36a4-4325-b520-e26b3dca309e" containerID="986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf" exitCode=0 Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.346752 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ldf4r" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="registry-server" containerID="cri-o://d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f" gracePeriod=2 Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.349368 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-d8msj" event={"ID":"fd3c60bd-b664-43b1-b669-3ef640959006","Type":"ContainerDied","Data":"2808165fe0dd2c32947b84019c507af32b1a337d587ed7291c288568222788ed"} Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.349727 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2808165fe0dd2c32947b84019c507af32b1a337d587ed7291c288568222788ed" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.349753 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"92a5651e-7175-45ca-a69b-dc19a11f2943","Type":"ContainerDied","Data":"bcdd2a68e5172c990298b42de9bed69bd8c00801bcd6c8e980fd5d7cc6fafff6"} Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.349776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63f863dc-36a4-4325-b520-e26b3dca309e","Type":"ContainerDied","Data":"986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf"} Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.376529 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-scripts\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.377055 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-dispersionconf\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.377313 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fd3c60bd-b664-43b1-b669-3ef640959006-etc-swift\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.377409 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-swiftconf\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.377474 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-combined-ca-bundle\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.377510 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq6g5\" (UniqueName: \"kubernetes.io/projected/fd3c60bd-b664-43b1-b669-3ef640959006-kube-api-access-jq6g5\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.377551 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-ring-data-devices\") pod \"fd3c60bd-b664-43b1-b669-3ef640959006\" (UID: \"fd3c60bd-b664-43b1-b669-3ef640959006\") " Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.378611 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.379434 4775 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.379482 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd3c60bd-b664-43b1-b669-3ef640959006-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.395114 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.400101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd3c60bd-b664-43b1-b669-3ef640959006-kube-api-access-jq6g5" (OuterVolumeSpecName: "kube-api-access-jq6g5") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "kube-api-access-jq6g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.424281 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.424537 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.424667 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.425598 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c257f11ab63c2232fe4ea03480103fc149650fe9c5d870044eeec21f9ccaaf75"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.425774 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://c257f11ab63c2232fe4ea03480103fc149650fe9c5d870044eeec21f9ccaaf75" gracePeriod=600 Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.434266 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.434325 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-scripts" (OuterVolumeSpecName: "scripts") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.437344 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd3c60bd-b664-43b1-b669-3ef640959006" (UID: "fd3c60bd-b664-43b1-b669-3ef640959006"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.489983 4775 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fd3c60bd-b664-43b1-b669-3ef640959006-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.490012 4775 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.490022 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.490031 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq6g5\" (UniqueName: \"kubernetes.io/projected/fd3c60bd-b664-43b1-b669-3ef640959006-kube-api-access-jq6g5\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.490040 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fd3c60bd-b664-43b1-b669-3ef640959006-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.490049 4775 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fd3c60bd-b664-43b1-b669-3ef640959006-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:51 crc kubenswrapper[4775]: I1126 06:40:51.907973 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.000930 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pftv\" (UniqueName: \"kubernetes.io/projected/0e3c05a3-6295-4d99-803a-a4495d2046fc-kube-api-access-8pftv\") pod \"0e3c05a3-6295-4d99-803a-a4495d2046fc\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.001003 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-catalog-content\") pod \"0e3c05a3-6295-4d99-803a-a4495d2046fc\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.001104 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-utilities\") pod \"0e3c05a3-6295-4d99-803a-a4495d2046fc\" (UID: \"0e3c05a3-6295-4d99-803a-a4495d2046fc\") " Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.002760 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-utilities" (OuterVolumeSpecName: "utilities") pod "0e3c05a3-6295-4d99-803a-a4495d2046fc" (UID: "0e3c05a3-6295-4d99-803a-a4495d2046fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.007605 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e3c05a3-6295-4d99-803a-a4495d2046fc-kube-api-access-8pftv" (OuterVolumeSpecName: "kube-api-access-8pftv") pod "0e3c05a3-6295-4d99-803a-a4495d2046fc" (UID: "0e3c05a3-6295-4d99-803a-a4495d2046fc"). InnerVolumeSpecName "kube-api-access-8pftv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.021188 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e3c05a3-6295-4d99-803a-a4495d2046fc" (UID: "0e3c05a3-6295-4d99-803a-a4495d2046fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.102431 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pftv\" (UniqueName: \"kubernetes.io/projected/0e3c05a3-6295-4d99-803a-a4495d2046fc-kube-api-access-8pftv\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.102464 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.102475 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e3c05a3-6295-4d99-803a-a4495d2046fc-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.355998 4775 generic.go:334] "Generic (PLEG): container finished" podID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerID="d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f" exitCode=0 Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.356041 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ldf4r" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.356061 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerDied","Data":"d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.356451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ldf4r" event={"ID":"0e3c05a3-6295-4d99-803a-a4495d2046fc","Type":"ContainerDied","Data":"976337157b3c33d5ba856ceebd84f2f538a087bf4b5e881bd8ef40251575bd0c"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.356475 4775 scope.go:117] "RemoveContainer" containerID="d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.359512 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="c257f11ab63c2232fe4ea03480103fc149650fe9c5d870044eeec21f9ccaaf75" exitCode=0 Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.359563 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"c257f11ab63c2232fe4ea03480103fc149650fe9c5d870044eeec21f9ccaaf75"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.359612 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"8e3a269c3e7b57838eb3de922a5f67cb80c1084327864aee098382792cdcca9f"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.361565 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"92a5651e-7175-45ca-a69b-dc19a11f2943","Type":"ContainerStarted","Data":"23f1b68e58af700cd8d63e29753ae5ef210dc630aedd23de681e823f78054454"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.361888 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.368433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bnfj2" event={"ID":"06776a0e-14fb-411c-b70d-861de9969bcd","Type":"ContainerStarted","Data":"ce0ab6e4d4886a95817154fa408d14227f06457a00a04fc5506c23eca7758e63"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.371069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63f863dc-36a4-4325-b520-e26b3dca309e","Type":"ContainerStarted","Data":"ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804"} Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.371299 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.376921 4775 scope.go:117] "RemoveContainer" containerID="747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.396990 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-bnfj2" podStartSLOduration=3.021382509 podStartE2EDuration="15.396968716s" podCreationTimestamp="2025-11-26 06:40:37 +0000 UTC" firstStartedPulling="2025-11-26 06:40:39.04980043 +0000 UTC m=+922.411104372" lastFinishedPulling="2025-11-26 06:40:51.425386617 +0000 UTC m=+934.786690579" observedRunningTime="2025-11-26 06:40:52.39523535 +0000 UTC m=+935.756539322" watchObservedRunningTime="2025-11-26 06:40:52.396968716 +0000 UTC m=+935.758272668" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.417068 4775 scope.go:117] "RemoveContainer" containerID="3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.441051 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.656235877 podStartE2EDuration="1m0.441029933s" podCreationTimestamp="2025-11-26 06:39:52 +0000 UTC" firstStartedPulling="2025-11-26 06:40:06.44943801 +0000 UTC m=+889.810741962" lastFinishedPulling="2025-11-26 06:40:16.234232066 +0000 UTC m=+899.595536018" observedRunningTime="2025-11-26 06:40:52.431244919 +0000 UTC m=+935.792548871" watchObservedRunningTime="2025-11-26 06:40:52.441029933 +0000 UTC m=+935.802333885" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.454329 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldf4r"] Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.459506 4775 scope.go:117] "RemoveContainer" containerID="d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f" Nov 26 06:40:52 crc kubenswrapper[4775]: E1126 06:40:52.460014 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f\": container with ID starting with d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f not found: ID does not exist" containerID="d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.460103 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f"} err="failed to get container status \"d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f\": rpc error: code = NotFound desc = could not find container \"d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f\": container with ID starting with d7728b7e64d69e43dc6731bd8f3cb7747fd03d6e65730fdefd03795dbbd26a5f not found: ID does not exist" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.460217 4775 scope.go:117] "RemoveContainer" containerID="747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.460260 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ldf4r"] Nov 26 06:40:52 crc kubenswrapper[4775]: E1126 06:40:52.463140 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036\": container with ID starting with 747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036 not found: ID does not exist" containerID="747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.463189 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036"} err="failed to get container status \"747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036\": rpc error: code = NotFound desc = could not find container \"747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036\": container with ID starting with 747c78a66d4881e033ff5c93f3f8f9c5226f4b90d7963291179550045776c036 not found: ID does not exist" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.463214 4775 scope.go:117] "RemoveContainer" containerID="3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f" Nov 26 06:40:52 crc kubenswrapper[4775]: E1126 06:40:52.463842 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f\": container with ID starting with 3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f not found: ID does not exist" containerID="3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.463865 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f"} err="failed to get container status \"3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f\": rpc error: code = NotFound desc = could not find container \"3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f\": container with ID starting with 3ce5df876af13913fcc16417da599153c07d66d88522ef0789e02e5a680b382f not found: ID does not exist" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.463883 4775 scope.go:117] "RemoveContainer" containerID="8685d35e684f235bf0161f8f3e1dacf8484e59a0d24a51ad51c239999335747e" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.482632 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.469991486 podStartE2EDuration="1m0.482614812s" podCreationTimestamp="2025-11-26 06:39:52 +0000 UTC" firstStartedPulling="2025-11-26 06:40:07.543063696 +0000 UTC m=+890.904367648" lastFinishedPulling="2025-11-26 06:40:16.555687022 +0000 UTC m=+899.916990974" observedRunningTime="2025-11-26 06:40:52.468930484 +0000 UTC m=+935.830234446" watchObservedRunningTime="2025-11-26 06:40:52.482614812 +0000 UTC m=+935.843918754" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.740272 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-k4mqj" podUID="d30382f9-28bd-4ab7-8fec-1f8932b3f804" containerName="ovn-controller" probeResult="failure" output=< Nov 26 06:40:52 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 26 06:40:52 crc kubenswrapper[4775]: > Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.771524 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:52 crc kubenswrapper[4775]: I1126 06:40:52.781336 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4qmsw" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020152 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-k4mqj-config-nbhzt"] Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.020783 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="registry-server" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020812 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="registry-server" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.020832 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3c60bd-b664-43b1-b669-3ef640959006" containerName="swift-ring-rebalance" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020843 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3c60bd-b664-43b1-b669-3ef640959006" containerName="swift-ring-rebalance" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.020867 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01958f53-a2a9-4783-9669-abd002d57a66" containerName="init" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020877 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="01958f53-a2a9-4783-9669-abd002d57a66" containerName="init" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.020899 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c54287b0-da13-43bc-9a94-d5fa303b885e" containerName="mariadb-database-create" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020909 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c54287b0-da13-43bc-9a94-d5fa303b885e" containerName="mariadb-database-create" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.020935 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="extract-utilities" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020945 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="extract-utilities" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.020963 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" containerName="mariadb-database-create" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.020973 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" containerName="mariadb-database-create" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.021001 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" containerName="mariadb-account-create-update" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021013 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" containerName="mariadb-account-create-update" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.021026 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="extract-content" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021038 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="extract-content" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.021051 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73081230-d192-4993-9434-9c4c6a5533f9" containerName="mariadb-account-create-update" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021062 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="73081230-d192-4993-9434-9c4c6a5533f9" containerName="mariadb-account-create-update" Nov 26 06:40:53 crc kubenswrapper[4775]: E1126 06:40:53.021086 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01958f53-a2a9-4783-9669-abd002d57a66" containerName="dnsmasq-dns" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021097 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="01958f53-a2a9-4783-9669-abd002d57a66" containerName="dnsmasq-dns" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021360 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c54287b0-da13-43bc-9a94-d5fa303b885e" containerName="mariadb-database-create" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021382 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="73081230-d192-4993-9434-9c4c6a5533f9" containerName="mariadb-account-create-update" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021400 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3c60bd-b664-43b1-b669-3ef640959006" containerName="swift-ring-rebalance" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021415 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" containerName="mariadb-database-create" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021433 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" containerName="mariadb-account-create-update" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021449 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="01958f53-a2a9-4783-9669-abd002d57a66" containerName="dnsmasq-dns" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.021468 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" containerName="registry-server" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.022346 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.027084 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.041516 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4mqj-config-nbhzt"] Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.119571 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-log-ovn\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.119638 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run-ovn\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.119725 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-additional-scripts\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.119742 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-scripts\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.119774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.119789 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fx92\" (UniqueName: \"kubernetes.io/projected/54824116-1e93-4db2-807d-d70f9ba98b53-kube-api-access-2fx92\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221327 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-log-ovn\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run-ovn\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-scripts\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221562 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-additional-scripts\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221607 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221631 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fx92\" (UniqueName: \"kubernetes.io/projected/54824116-1e93-4db2-807d-d70f9ba98b53-kube-api-access-2fx92\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221652 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-log-ovn\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221657 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run-ovn\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.221781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.222372 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-additional-scripts\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.223657 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-scripts\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.244448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fx92\" (UniqueName: \"kubernetes.io/projected/54824116-1e93-4db2-807d-d70f9ba98b53-kube-api-access-2fx92\") pod \"ovn-controller-k4mqj-config-nbhzt\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.339612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.340041 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e3c05a3-6295-4d99-803a-a4495d2046fc" path="/var/lib/kubelet/pods/0e3c05a3-6295-4d99-803a-a4495d2046fc/volumes" Nov 26 06:40:53 crc kubenswrapper[4775]: I1126 06:40:53.830370 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-k4mqj-config-nbhzt"] Nov 26 06:40:54 crc kubenswrapper[4775]: I1126 06:40:54.418203 4775 generic.go:334] "Generic (PLEG): container finished" podID="54824116-1e93-4db2-807d-d70f9ba98b53" containerID="b5ba65b78191c56fd2f5ab03c97b3736920b1b6922f7422278b37c8933da576c" exitCode=0 Nov 26 06:40:54 crc kubenswrapper[4775]: I1126 06:40:54.418240 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4mqj-config-nbhzt" event={"ID":"54824116-1e93-4db2-807d-d70f9ba98b53","Type":"ContainerDied","Data":"b5ba65b78191c56fd2f5ab03c97b3736920b1b6922f7422278b37c8933da576c"} Nov 26 06:40:54 crc kubenswrapper[4775]: I1126 06:40:54.418559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4mqj-config-nbhzt" event={"ID":"54824116-1e93-4db2-807d-d70f9ba98b53","Type":"ContainerStarted","Data":"b7492fe8aa28e8ed5d0e01bc4d27be5c72cbcf14a9ab3af1dc761b8cbe5c2e04"} Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.796456 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.865842 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-log-ovn\") pod \"54824116-1e93-4db2-807d-d70f9ba98b53\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.865928 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run\") pod \"54824116-1e93-4db2-807d-d70f9ba98b53\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.865962 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "54824116-1e93-4db2-807d-d70f9ba98b53" (UID: "54824116-1e93-4db2-807d-d70f9ba98b53"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.865988 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-scripts\") pod \"54824116-1e93-4db2-807d-d70f9ba98b53\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.865989 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run" (OuterVolumeSpecName: "var-run") pod "54824116-1e93-4db2-807d-d70f9ba98b53" (UID: "54824116-1e93-4db2-807d-d70f9ba98b53"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866072 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run-ovn\") pod \"54824116-1e93-4db2-807d-d70f9ba98b53\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866186 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-additional-scripts\") pod \"54824116-1e93-4db2-807d-d70f9ba98b53\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866181 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "54824116-1e93-4db2-807d-d70f9ba98b53" (UID: "54824116-1e93-4db2-807d-d70f9ba98b53"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866244 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fx92\" (UniqueName: \"kubernetes.io/projected/54824116-1e93-4db2-807d-d70f9ba98b53-kube-api-access-2fx92\") pod \"54824116-1e93-4db2-807d-d70f9ba98b53\" (UID: \"54824116-1e93-4db2-807d-d70f9ba98b53\") " Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866704 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "54824116-1e93-4db2-807d-d70f9ba98b53" (UID: "54824116-1e93-4db2-807d-d70f9ba98b53"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866849 4775 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866864 4775 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866872 4775 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/54824116-1e93-4db2-807d-d70f9ba98b53-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.866880 4775 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.867177 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-scripts" (OuterVolumeSpecName: "scripts") pod "54824116-1e93-4db2-807d-d70f9ba98b53" (UID: "54824116-1e93-4db2-807d-d70f9ba98b53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.882862 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54824116-1e93-4db2-807d-d70f9ba98b53-kube-api-access-2fx92" (OuterVolumeSpecName: "kube-api-access-2fx92") pod "54824116-1e93-4db2-807d-d70f9ba98b53" (UID: "54824116-1e93-4db2-807d-d70f9ba98b53"). InnerVolumeSpecName "kube-api-access-2fx92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.967825 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54824116-1e93-4db2-807d-d70f9ba98b53-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:55 crc kubenswrapper[4775]: I1126 06:40:55.968257 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fx92\" (UniqueName: \"kubernetes.io/projected/54824116-1e93-4db2-807d-d70f9ba98b53-kube-api-access-2fx92\") on node \"crc\" DevicePath \"\"" Nov 26 06:40:56 crc kubenswrapper[4775]: I1126 06:40:56.444375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-k4mqj-config-nbhzt" event={"ID":"54824116-1e93-4db2-807d-d70f9ba98b53","Type":"ContainerDied","Data":"b7492fe8aa28e8ed5d0e01bc4d27be5c72cbcf14a9ab3af1dc761b8cbe5c2e04"} Nov 26 06:40:56 crc kubenswrapper[4775]: I1126 06:40:56.444444 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7492fe8aa28e8ed5d0e01bc4d27be5c72cbcf14a9ab3af1dc761b8cbe5c2e04" Nov 26 06:40:56 crc kubenswrapper[4775]: I1126 06:40:56.444563 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-k4mqj-config-nbhzt" Nov 26 06:40:56 crc kubenswrapper[4775]: I1126 06:40:56.942436 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-k4mqj-config-nbhzt"] Nov 26 06:40:56 crc kubenswrapper[4775]: I1126 06:40:56.950787 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-k4mqj-config-nbhzt"] Nov 26 06:40:57 crc kubenswrapper[4775]: I1126 06:40:57.339983 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54824116-1e93-4db2-807d-d70f9ba98b53" path="/var/lib/kubelet/pods/54824116-1e93-4db2-807d-d70f9ba98b53/volumes" Nov 26 06:40:57 crc kubenswrapper[4775]: I1126 06:40:57.745384 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-k4mqj" Nov 26 06:40:58 crc kubenswrapper[4775]: I1126 06:40:58.467781 4775 generic.go:334] "Generic (PLEG): container finished" podID="06776a0e-14fb-411c-b70d-861de9969bcd" containerID="ce0ab6e4d4886a95817154fa408d14227f06457a00a04fc5506c23eca7758e63" exitCode=0 Nov 26 06:40:58 crc kubenswrapper[4775]: I1126 06:40:58.467841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bnfj2" event={"ID":"06776a0e-14fb-411c-b70d-861de9969bcd","Type":"ContainerDied","Data":"ce0ab6e4d4886a95817154fa408d14227f06457a00a04fc5506c23eca7758e63"} Nov 26 06:40:59 crc kubenswrapper[4775]: I1126 06:40:59.953255 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bnfj2" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.047327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-config-data\") pod \"06776a0e-14fb-411c-b70d-861de9969bcd\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.047522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-db-sync-config-data\") pod \"06776a0e-14fb-411c-b70d-861de9969bcd\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.047581 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldkxb\" (UniqueName: \"kubernetes.io/projected/06776a0e-14fb-411c-b70d-861de9969bcd-kube-api-access-ldkxb\") pod \"06776a0e-14fb-411c-b70d-861de9969bcd\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.047640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-combined-ca-bundle\") pod \"06776a0e-14fb-411c-b70d-861de9969bcd\" (UID: \"06776a0e-14fb-411c-b70d-861de9969bcd\") " Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.060087 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06776a0e-14fb-411c-b70d-861de9969bcd-kube-api-access-ldkxb" (OuterVolumeSpecName: "kube-api-access-ldkxb") pod "06776a0e-14fb-411c-b70d-861de9969bcd" (UID: "06776a0e-14fb-411c-b70d-861de9969bcd"). InnerVolumeSpecName "kube-api-access-ldkxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.062844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "06776a0e-14fb-411c-b70d-861de9969bcd" (UID: "06776a0e-14fb-411c-b70d-861de9969bcd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.094420 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06776a0e-14fb-411c-b70d-861de9969bcd" (UID: "06776a0e-14fb-411c-b70d-861de9969bcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.105359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-config-data" (OuterVolumeSpecName: "config-data") pod "06776a0e-14fb-411c-b70d-861de9969bcd" (UID: "06776a0e-14fb-411c-b70d-861de9969bcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.150528 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.150580 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldkxb\" (UniqueName: \"kubernetes.io/projected/06776a0e-14fb-411c-b70d-861de9969bcd-kube-api-access-ldkxb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.150599 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.150610 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06776a0e-14fb-411c-b70d-861de9969bcd-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.490182 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bnfj2" event={"ID":"06776a0e-14fb-411c-b70d-861de9969bcd","Type":"ContainerDied","Data":"4290294b61fed7ea376a6005f60be7a8f5f314b58028aabf6a9cc63f2d377655"} Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.490224 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4290294b61fed7ea376a6005f60be7a8f5f314b58028aabf6a9cc63f2d377655" Nov 26 06:41:00 crc kubenswrapper[4775]: I1126 06:41:00.490286 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bnfj2" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.015497 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jggzv"] Nov 26 06:41:01 crc kubenswrapper[4775]: E1126 06:41:01.016094 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06776a0e-14fb-411c-b70d-861de9969bcd" containerName="glance-db-sync" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.016109 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="06776a0e-14fb-411c-b70d-861de9969bcd" containerName="glance-db-sync" Nov 26 06:41:01 crc kubenswrapper[4775]: E1126 06:41:01.016138 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54824116-1e93-4db2-807d-d70f9ba98b53" containerName="ovn-config" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.016144 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="54824116-1e93-4db2-807d-d70f9ba98b53" containerName="ovn-config" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.016280 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="06776a0e-14fb-411c-b70d-861de9969bcd" containerName="glance-db-sync" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.016293 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="54824116-1e93-4db2-807d-d70f9ba98b53" containerName="ovn-config" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.017129 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.034300 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jggzv"] Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.166744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.166793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.167118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx4nr\" (UniqueName: \"kubernetes.io/projected/ed38647f-f4ed-4993-83f1-f7f34255e9b7-kube-api-access-lx4nr\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.167183 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.167266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-config\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.268209 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.268272 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.269448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.269459 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.269690 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx4nr\" (UniqueName: \"kubernetes.io/projected/ed38647f-f4ed-4993-83f1-f7f34255e9b7-kube-api-access-lx4nr\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.269729 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.269769 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-config\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.270344 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-config\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.270841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.294512 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx4nr\" (UniqueName: \"kubernetes.io/projected/ed38647f-f4ed-4993-83f1-f7f34255e9b7-kube-api-access-lx4nr\") pod \"dnsmasq-dns-5b946c75cc-jggzv\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.330382 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:01 crc kubenswrapper[4775]: I1126 06:41:01.832918 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jggzv"] Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.085677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.092805 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/38df1b80-132c-467f-969c-d3798dd1f7d9-etc-swift\") pod \"swift-storage-0\" (UID: \"38df1b80-132c-467f-969c-d3798dd1f7d9\") " pod="openstack/swift-storage-0" Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.297744 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.507540 4775 generic.go:334] "Generic (PLEG): container finished" podID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerID="42062ad7ec7b4257800732aef18d6bd9d02d0261bc387d67ac6eecda9946f547" exitCode=0 Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.507605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" event={"ID":"ed38647f-f4ed-4993-83f1-f7f34255e9b7","Type":"ContainerDied","Data":"42062ad7ec7b4257800732aef18d6bd9d02d0261bc387d67ac6eecda9946f547"} Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.507905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" event={"ID":"ed38647f-f4ed-4993-83f1-f7f34255e9b7","Type":"ContainerStarted","Data":"ecd9b4f002cecb047982e56333b71e2897238a0db7e9427485d008ff060ec81d"} Nov 26 06:41:02 crc kubenswrapper[4775]: I1126 06:41:02.853402 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 26 06:41:02 crc kubenswrapper[4775]: W1126 06:41:02.863773 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38df1b80_132c_467f_969c_d3798dd1f7d9.slice/crio-c8d83c1d4bed4a90055283c67b6b32a9e33c78c2f0bea98286e4dc90fdee00b7 WatchSource:0}: Error finding container c8d83c1d4bed4a90055283c67b6b32a9e33c78c2f0bea98286e4dc90fdee00b7: Status 404 returned error can't find the container with id c8d83c1d4bed4a90055283c67b6b32a9e33c78c2f0bea98286e4dc90fdee00b7 Nov 26 06:41:04 crc kubenswrapper[4775]: I1126 06:41:04.141294 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:41:04 crc kubenswrapper[4775]: I1126 06:41:04.141660 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 26 06:41:04 crc kubenswrapper[4775]: I1126 06:41:04.141675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" event={"ID":"ed38647f-f4ed-4993-83f1-f7f34255e9b7","Type":"ContainerStarted","Data":"e8aebe0fab1e656480dc0057b764996044e1fa9b766e37f0b917ec6ebbed7bcc"} Nov 26 06:41:04 crc kubenswrapper[4775]: I1126 06:41:04.144029 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"c8d83c1d4bed4a90055283c67b6b32a9e33c78c2f0bea98286e4dc90fdee00b7"} Nov 26 06:41:04 crc kubenswrapper[4775]: I1126 06:41:04.162909 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" podStartSLOduration=4.162890869 podStartE2EDuration="4.162890869s" podCreationTimestamp="2025-11-26 06:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:04.16144532 +0000 UTC m=+947.522749312" watchObservedRunningTime="2025-11-26 06:41:04.162890869 +0000 UTC m=+947.524194831" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.153135 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.780440 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-fkk8f"] Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.781596 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.810164 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fkk8f"] Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.893173 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-w8tnl"] Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.894240 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.904630 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-w8tnl"] Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.923788 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-dd94-account-create-update-7zjm6"] Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.925179 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.931018 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.946210 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dd94-account-create-update-7zjm6"] Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.951414 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr462\" (UniqueName: \"kubernetes.io/projected/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-kube-api-access-xr462\") pod \"barbican-db-create-fkk8f\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:05 crc kubenswrapper[4775]: I1126 06:41:05.951480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-operator-scripts\") pod \"barbican-db-create-fkk8f\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.018746 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c371-account-create-update-jt4mk"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.019884 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.024029 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.036583 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c371-account-create-update-jt4mk"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.053063 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4pdb\" (UniqueName: \"kubernetes.io/projected/8d56ad5c-f3dc-4d36-862b-a68529cffacb-kube-api-access-k4pdb\") pod \"cinder-dd94-account-create-update-7zjm6\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.053119 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-operator-scripts\") pod \"barbican-db-create-fkk8f\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.053166 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d56ad5c-f3dc-4d36-862b-a68529cffacb-operator-scripts\") pod \"cinder-dd94-account-create-update-7zjm6\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.053226 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8c4\" (UniqueName: \"kubernetes.io/projected/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-kube-api-access-kk8c4\") pod \"cinder-db-create-w8tnl\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.053270 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-operator-scripts\") pod \"cinder-db-create-w8tnl\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.053294 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr462\" (UniqueName: \"kubernetes.io/projected/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-kube-api-access-xr462\") pod \"barbican-db-create-fkk8f\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.054226 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-operator-scripts\") pod \"barbican-db-create-fkk8f\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.077753 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-j4rs9"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.079201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.093812 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr462\" (UniqueName: \"kubernetes.io/projected/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-kube-api-access-xr462\") pod \"barbican-db-create-fkk8f\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.100538 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.101410 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-j4rs9"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.158311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c651662-819c-4691-9632-5948a2b3d88c-operator-scripts\") pod \"barbican-c371-account-create-update-jt4mk\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.159340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67rl5\" (UniqueName: \"kubernetes.io/projected/9c651662-819c-4691-9632-5948a2b3d88c-kube-api-access-67rl5\") pod \"barbican-c371-account-create-update-jt4mk\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.159478 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8c4\" (UniqueName: \"kubernetes.io/projected/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-kube-api-access-kk8c4\") pod \"cinder-db-create-w8tnl\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.159595 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-operator-scripts\") pod \"cinder-db-create-w8tnl\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.159740 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4pdb\" (UniqueName: \"kubernetes.io/projected/8d56ad5c-f3dc-4d36-862b-a68529cffacb-kube-api-access-k4pdb\") pod \"cinder-dd94-account-create-update-7zjm6\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.159883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d56ad5c-f3dc-4d36-862b-a68529cffacb-operator-scripts\") pod \"cinder-dd94-account-create-update-7zjm6\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.160857 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d56ad5c-f3dc-4d36-862b-a68529cffacb-operator-scripts\") pod \"cinder-dd94-account-create-update-7zjm6\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.161504 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-operator-scripts\") pod \"cinder-db-create-w8tnl\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.162462 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-7kcv8"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.186295 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7kcv8"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.186387 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.195301 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.195601 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.195363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8c4\" (UniqueName: \"kubernetes.io/projected/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-kube-api-access-kk8c4\") pod \"cinder-db-create-w8tnl\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.196501 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.196786 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jvmv4" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.198663 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4pdb\" (UniqueName: \"kubernetes.io/projected/8d56ad5c-f3dc-4d36-862b-a68529cffacb-kube-api-access-k4pdb\") pod \"cinder-dd94-account-create-update-7zjm6\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.211063 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.255550 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.261304 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c651662-819c-4691-9632-5948a2b3d88c-operator-scripts\") pod \"barbican-c371-account-create-update-jt4mk\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.261423 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwjxj\" (UniqueName: \"kubernetes.io/projected/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-kube-api-access-jwjxj\") pod \"neutron-db-create-j4rs9\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.261584 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67rl5\" (UniqueName: \"kubernetes.io/projected/9c651662-819c-4691-9632-5948a2b3d88c-kube-api-access-67rl5\") pod \"barbican-c371-account-create-update-jt4mk\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.261752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-operator-scripts\") pod \"neutron-db-create-j4rs9\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.262841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c651662-819c-4691-9632-5948a2b3d88c-operator-scripts\") pod \"barbican-c371-account-create-update-jt4mk\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.314097 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c40d-account-create-update-snvdn"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.314689 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67rl5\" (UniqueName: \"kubernetes.io/projected/9c651662-819c-4691-9632-5948a2b3d88c-kube-api-access-67rl5\") pod \"barbican-c371-account-create-update-jt4mk\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.316707 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.320411 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.325735 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c40d-account-create-update-snvdn"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.344150 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.364003 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-combined-ca-bundle\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.364368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwjxj\" (UniqueName: \"kubernetes.io/projected/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-kube-api-access-jwjxj\") pod \"neutron-db-create-j4rs9\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.364816 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-operator-scripts\") pod \"neutron-db-create-j4rs9\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.364918 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-config-data\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.365040 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqtzf\" (UniqueName: \"kubernetes.io/projected/1ea8f43d-f87b-4f26-9d09-465c46a9e853-kube-api-access-xqtzf\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.367095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-operator-scripts\") pod \"neutron-db-create-j4rs9\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.389405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwjxj\" (UniqueName: \"kubernetes.io/projected/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-kube-api-access-jwjxj\") pod \"neutron-db-create-j4rs9\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.428151 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.467929 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-config-data\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.468041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqtzf\" (UniqueName: \"kubernetes.io/projected/1ea8f43d-f87b-4f26-9d09-465c46a9e853-kube-api-access-xqtzf\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.468074 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-combined-ca-bundle\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.468134 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f96dh\" (UniqueName: \"kubernetes.io/projected/229de831-b61a-4cee-8005-5a53eede49a1-kube-api-access-f96dh\") pod \"neutron-c40d-account-create-update-snvdn\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.468189 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229de831-b61a-4cee-8005-5a53eede49a1-operator-scripts\") pod \"neutron-c40d-account-create-update-snvdn\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.475620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-combined-ca-bundle\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.476098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-config-data\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.487039 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqtzf\" (UniqueName: \"kubernetes.io/projected/1ea8f43d-f87b-4f26-9d09-465c46a9e853-kube-api-access-xqtzf\") pod \"keystone-db-sync-7kcv8\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.569943 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f96dh\" (UniqueName: \"kubernetes.io/projected/229de831-b61a-4cee-8005-5a53eede49a1-kube-api-access-f96dh\") pod \"neutron-c40d-account-create-update-snvdn\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.570337 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229de831-b61a-4cee-8005-5a53eede49a1-operator-scripts\") pod \"neutron-c40d-account-create-update-snvdn\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.571111 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229de831-b61a-4cee-8005-5a53eede49a1-operator-scripts\") pod \"neutron-c40d-account-create-update-snvdn\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.585286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f96dh\" (UniqueName: \"kubernetes.io/projected/229de831-b61a-4cee-8005-5a53eede49a1-kube-api-access-f96dh\") pod \"neutron-c40d-account-create-update-snvdn\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.620547 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.634397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.769253 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-fkk8f"] Nov 26 06:41:06 crc kubenswrapper[4775]: W1126 06:41:06.788248 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac40cdd2_dadd_43ee_bf2f_d6df962d5d52.slice/crio-d015db5968dae648e3070b7199cfb913b64a8670ccd52f0d861feed81cf196f6 WatchSource:0}: Error finding container d015db5968dae648e3070b7199cfb913b64a8670ccd52f0d861feed81cf196f6: Status 404 returned error can't find the container with id d015db5968dae648e3070b7199cfb913b64a8670ccd52f0d861feed81cf196f6 Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.829165 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-w8tnl"] Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.837999 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dd94-account-create-update-7zjm6"] Nov 26 06:41:06 crc kubenswrapper[4775]: W1126 06:41:06.892860 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d56ad5c_f3dc_4d36_862b_a68529cffacb.slice/crio-ee1203cd60e8e219f95c6e490eded27ff81b9ea7f7ba1c49a3241894e43bb25a WatchSource:0}: Error finding container ee1203cd60e8e219f95c6e490eded27ff81b9ea7f7ba1c49a3241894e43bb25a: Status 404 returned error can't find the container with id ee1203cd60e8e219f95c6e490eded27ff81b9ea7f7ba1c49a3241894e43bb25a Nov 26 06:41:06 crc kubenswrapper[4775]: I1126 06:41:06.958105 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c371-account-create-update-jt4mk"] Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.004542 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-j4rs9"] Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.127573 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-7kcv8"] Nov 26 06:41:07 crc kubenswrapper[4775]: W1126 06:41:07.129628 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ea8f43d_f87b_4f26_9d09_465c46a9e853.slice/crio-03ebae7f67c5f92384fbccdda6dec0b8d607fbf2052e01357db700c5828cc0f3 WatchSource:0}: Error finding container 03ebae7f67c5f92384fbccdda6dec0b8d607fbf2052e01357db700c5828cc0f3: Status 404 returned error can't find the container with id 03ebae7f67c5f92384fbccdda6dec0b8d607fbf2052e01357db700c5828cc0f3 Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.199905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7kcv8" event={"ID":"1ea8f43d-f87b-4f26-9d09-465c46a9e853","Type":"ContainerStarted","Data":"03ebae7f67c5f92384fbccdda6dec0b8d607fbf2052e01357db700c5828cc0f3"} Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.200824 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-j4rs9" event={"ID":"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5","Type":"ContainerStarted","Data":"aa6e27dfc901aaeab8ca6623f3d44b3ed62da84b5f7df9d6c802e3f4e63de004"} Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.201681 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-w8tnl" event={"ID":"a7e342a0-cce3-4690-93ef-f1afd6f8e44a","Type":"ContainerStarted","Data":"c56b36128f537e202d4247f8584d0acc8df3d38417fc0ce3676faf0ac3e20cfe"} Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.202660 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fkk8f" event={"ID":"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52","Type":"ContainerStarted","Data":"d015db5968dae648e3070b7199cfb913b64a8670ccd52f0d861feed81cf196f6"} Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.203943 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c371-account-create-update-jt4mk" event={"ID":"9c651662-819c-4691-9632-5948a2b3d88c","Type":"ContainerStarted","Data":"14361cf30e37e2df7a982c2c2b49e6e77d7737c66befe5d1931e879ac071ad85"} Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.207219 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dd94-account-create-update-7zjm6" event={"ID":"8d56ad5c-f3dc-4d36-862b-a68529cffacb","Type":"ContainerStarted","Data":"ee1203cd60e8e219f95c6e490eded27ff81b9ea7f7ba1c49a3241894e43bb25a"} Nov 26 06:41:07 crc kubenswrapper[4775]: W1126 06:41:07.256953 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229de831_b61a_4cee_8005_5a53eede49a1.slice/crio-66a0e1fffc833a05c577b309450f6aa7c4983569fb0702a6fa029cfc32471656 WatchSource:0}: Error finding container 66a0e1fffc833a05c577b309450f6aa7c4983569fb0702a6fa029cfc32471656: Status 404 returned error can't find the container with id 66a0e1fffc833a05c577b309450f6aa7c4983569fb0702a6fa029cfc32471656 Nov 26 06:41:07 crc kubenswrapper[4775]: I1126 06:41:07.263789 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c40d-account-create-update-snvdn"] Nov 26 06:41:08 crc kubenswrapper[4775]: I1126 06:41:08.216889 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c40d-account-create-update-snvdn" event={"ID":"229de831-b61a-4cee-8005-5a53eede49a1","Type":"ContainerStarted","Data":"66a0e1fffc833a05c577b309450f6aa7c4983569fb0702a6fa029cfc32471656"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.227918 4775 generic.go:334] "Generic (PLEG): container finished" podID="229de831-b61a-4cee-8005-5a53eede49a1" containerID="6e0c23254a48848983f3df52b3e0dc6527463727d8d184c9322ed8149fef90ff" exitCode=0 Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.227994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c40d-account-create-update-snvdn" event={"ID":"229de831-b61a-4cee-8005-5a53eede49a1","Type":"ContainerDied","Data":"6e0c23254a48848983f3df52b3e0dc6527463727d8d184c9322ed8149fef90ff"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.235644 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"49e4bb1158d56ad78761ff94b15393a586eb46bf4b9fa831744918dee0de9e3f"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.237542 4775 generic.go:334] "Generic (PLEG): container finished" podID="759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" containerID="53b8e58d4dccb836949653f67bce2bf8bad8e30359f230e26554bf8609f65956" exitCode=0 Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.237615 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-j4rs9" event={"ID":"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5","Type":"ContainerDied","Data":"53b8e58d4dccb836949653f67bce2bf8bad8e30359f230e26554bf8609f65956"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.239106 4775 generic.go:334] "Generic (PLEG): container finished" podID="a7e342a0-cce3-4690-93ef-f1afd6f8e44a" containerID="6b41a0b963dfac26b8b087dfba156282ed18e588d8998ac10ded3a2f100c4d6d" exitCode=0 Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.239161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-w8tnl" event={"ID":"a7e342a0-cce3-4690-93ef-f1afd6f8e44a","Type":"ContainerDied","Data":"6b41a0b963dfac26b8b087dfba156282ed18e588d8998ac10ded3a2f100c4d6d"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.240773 4775 generic.go:334] "Generic (PLEG): container finished" podID="ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" containerID="058f3a49e4a2725fc75468f5bd01076a3782b831bc88a996b92b3f5c6350b24d" exitCode=0 Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.240876 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fkk8f" event={"ID":"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52","Type":"ContainerDied","Data":"058f3a49e4a2725fc75468f5bd01076a3782b831bc88a996b92b3f5c6350b24d"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.251784 4775 generic.go:334] "Generic (PLEG): container finished" podID="9c651662-819c-4691-9632-5948a2b3d88c" containerID="7c4e4a3df95422486a05e3aa000604a4931381c03330e4bfbaa59a7484d49f93" exitCode=0 Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.251825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c371-account-create-update-jt4mk" event={"ID":"9c651662-819c-4691-9632-5948a2b3d88c","Type":"ContainerDied","Data":"7c4e4a3df95422486a05e3aa000604a4931381c03330e4bfbaa59a7484d49f93"} Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.253574 4775 generic.go:334] "Generic (PLEG): container finished" podID="8d56ad5c-f3dc-4d36-862b-a68529cffacb" containerID="b698c78b7fb72307ce47dcf4e5ae0aeb354e3604c1148cc1a8353503afb6394e" exitCode=0 Nov 26 06:41:09 crc kubenswrapper[4775]: I1126 06:41:09.253605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dd94-account-create-update-7zjm6" event={"ID":"8d56ad5c-f3dc-4d36-862b-a68529cffacb","Type":"ContainerDied","Data":"b698c78b7fb72307ce47dcf4e5ae0aeb354e3604c1148cc1a8353503afb6394e"} Nov 26 06:41:10 crc kubenswrapper[4775]: I1126 06:41:10.262772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"52044c462c0768c76bd10ef45ba8cb0684aa74e916dfe8afcef03afa1b9399ad"} Nov 26 06:41:10 crc kubenswrapper[4775]: I1126 06:41:10.263144 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"1f1bc6d8333209c80bf4fceec18a0c023bc6d2f93dff597bdea969d2db61d4be"} Nov 26 06:41:10 crc kubenswrapper[4775]: I1126 06:41:10.263160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"300160475cc30c9af1addcc12ec4f35f0f3ef4d4be19dc8b2798ba5944d8e31e"} Nov 26 06:41:11 crc kubenswrapper[4775]: I1126 06:41:11.337355 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:11 crc kubenswrapper[4775]: I1126 06:41:11.419111 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2nsvj"] Nov 26 06:41:11 crc kubenswrapper[4775]: I1126 06:41:11.419411 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-2nsvj" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerName="dnsmasq-dns" containerID="cri-o://2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296" gracePeriod=10 Nov 26 06:41:11 crc kubenswrapper[4775]: E1126 06:41:11.648952 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e5e0f59_7e39_43c9_b9d4_fe8dd69ec635.slice/crio-conmon-2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e5e0f59_7e39_43c9_b9d4_fe8dd69ec635.slice/crio-2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296.scope\": RecentStats: unable to find data in memory cache]" Nov 26 06:41:12 crc kubenswrapper[4775]: I1126 06:41:12.284075 4775 generic.go:334] "Generic (PLEG): container finished" podID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerID="2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296" exitCode=0 Nov 26 06:41:12 crc kubenswrapper[4775]: I1126 06:41:12.284383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2nsvj" event={"ID":"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635","Type":"ContainerDied","Data":"2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296"} Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.715639 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.744499 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.756301 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.823327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk8c4\" (UniqueName: \"kubernetes.io/projected/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-kube-api-access-kk8c4\") pod \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.823408 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-operator-scripts\") pod \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\" (UID: \"a7e342a0-cce3-4690-93ef-f1afd6f8e44a\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.823479 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-operator-scripts\") pod \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.823502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwjxj\" (UniqueName: \"kubernetes.io/projected/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-kube-api-access-jwjxj\") pod \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\" (UID: \"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.823981 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.824422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" (UID: "759a51cc-5fd3-4c7d-bfd8-072222f1e2e5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.825275 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7e342a0-cce3-4690-93ef-f1afd6f8e44a" (UID: "a7e342a0-cce3-4690-93ef-f1afd6f8e44a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.827940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-kube-api-access-jwjxj" (OuterVolumeSpecName: "kube-api-access-jwjxj") pod "759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" (UID: "759a51cc-5fd3-4c7d-bfd8-072222f1e2e5"). InnerVolumeSpecName "kube-api-access-jwjxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.827983 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-kube-api-access-kk8c4" (OuterVolumeSpecName: "kube-api-access-kk8c4") pod "a7e342a0-cce3-4690-93ef-f1afd6f8e44a" (UID: "a7e342a0-cce3-4690-93ef-f1afd6f8e44a"). InnerVolumeSpecName "kube-api-access-kk8c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.847815 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.855906 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.864103 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.924905 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr462\" (UniqueName: \"kubernetes.io/projected/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-kube-api-access-xr462\") pod \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.924965 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f96dh\" (UniqueName: \"kubernetes.io/projected/229de831-b61a-4cee-8005-5a53eede49a1-kube-api-access-f96dh\") pod \"229de831-b61a-4cee-8005-5a53eede49a1\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925003 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67rl5\" (UniqueName: \"kubernetes.io/projected/9c651662-819c-4691-9632-5948a2b3d88c-kube-api-access-67rl5\") pod \"9c651662-819c-4691-9632-5948a2b3d88c\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925088 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-operator-scripts\") pod \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\" (UID: \"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925178 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229de831-b61a-4cee-8005-5a53eede49a1-operator-scripts\") pod \"229de831-b61a-4cee-8005-5a53eede49a1\" (UID: \"229de831-b61a-4cee-8005-5a53eede49a1\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925219 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c651662-819c-4691-9632-5948a2b3d88c-operator-scripts\") pod \"9c651662-819c-4691-9632-5948a2b3d88c\" (UID: \"9c651662-819c-4691-9632-5948a2b3d88c\") " Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925561 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925573 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwjxj\" (UniqueName: \"kubernetes.io/projected/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5-kube-api-access-jwjxj\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925583 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk8c4\" (UniqueName: \"kubernetes.io/projected/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-kube-api-access-kk8c4\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.925592 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7e342a0-cce3-4690-93ef-f1afd6f8e44a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.926779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" (UID: "ac40cdd2-dadd-43ee-bf2f-d6df962d5d52"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.926813 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229de831-b61a-4cee-8005-5a53eede49a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "229de831-b61a-4cee-8005-5a53eede49a1" (UID: "229de831-b61a-4cee-8005-5a53eede49a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.926929 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c651662-819c-4691-9632-5948a2b3d88c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c651662-819c-4691-9632-5948a2b3d88c" (UID: "9c651662-819c-4691-9632-5948a2b3d88c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.934797 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229de831-b61a-4cee-8005-5a53eede49a1-kube-api-access-f96dh" (OuterVolumeSpecName: "kube-api-access-f96dh") pod "229de831-b61a-4cee-8005-5a53eede49a1" (UID: "229de831-b61a-4cee-8005-5a53eede49a1"). InnerVolumeSpecName "kube-api-access-f96dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.938274 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c651662-819c-4691-9632-5948a2b3d88c-kube-api-access-67rl5" (OuterVolumeSpecName: "kube-api-access-67rl5") pod "9c651662-819c-4691-9632-5948a2b3d88c" (UID: "9c651662-819c-4691-9632-5948a2b3d88c"). InnerVolumeSpecName "kube-api-access-67rl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:13 crc kubenswrapper[4775]: I1126 06:41:13.939959 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-kube-api-access-xr462" (OuterVolumeSpecName: "kube-api-access-xr462") pod "ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" (UID: "ac40cdd2-dadd-43ee-bf2f-d6df962d5d52"). InnerVolumeSpecName "kube-api-access-xr462". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026356 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4pdb\" (UniqueName: \"kubernetes.io/projected/8d56ad5c-f3dc-4d36-862b-a68529cffacb-kube-api-access-k4pdb\") pod \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-sb\") pod \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026443 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpqtf\" (UniqueName: \"kubernetes.io/projected/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-kube-api-access-cpqtf\") pod \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026473 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-nb\") pod \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026501 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d56ad5c-f3dc-4d36-862b-a68529cffacb-operator-scripts\") pod \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\" (UID: \"8d56ad5c-f3dc-4d36-862b-a68529cffacb\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026563 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-config\") pod \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026648 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-dns-svc\") pod \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\" (UID: \"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635\") " Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026949 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr462\" (UniqueName: \"kubernetes.io/projected/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-kube-api-access-xr462\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026960 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f96dh\" (UniqueName: \"kubernetes.io/projected/229de831-b61a-4cee-8005-5a53eede49a1-kube-api-access-f96dh\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026973 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67rl5\" (UniqueName: \"kubernetes.io/projected/9c651662-819c-4691-9632-5948a2b3d88c-kube-api-access-67rl5\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.026990 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.027003 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229de831-b61a-4cee-8005-5a53eede49a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.027011 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c651662-819c-4691-9632-5948a2b3d88c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.027396 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d56ad5c-f3dc-4d36-862b-a68529cffacb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d56ad5c-f3dc-4d36-862b-a68529cffacb" (UID: "8d56ad5c-f3dc-4d36-862b-a68529cffacb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.030354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d56ad5c-f3dc-4d36-862b-a68529cffacb-kube-api-access-k4pdb" (OuterVolumeSpecName: "kube-api-access-k4pdb") pod "8d56ad5c-f3dc-4d36-862b-a68529cffacb" (UID: "8d56ad5c-f3dc-4d36-862b-a68529cffacb"). InnerVolumeSpecName "kube-api-access-k4pdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.032172 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-kube-api-access-cpqtf" (OuterVolumeSpecName: "kube-api-access-cpqtf") pod "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" (UID: "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635"). InnerVolumeSpecName "kube-api-access-cpqtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.064327 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" (UID: "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.072408 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" (UID: "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.073354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-config" (OuterVolumeSpecName: "config") pod "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" (UID: "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.074494 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" (UID: "9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128351 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128384 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpqtf\" (UniqueName: \"kubernetes.io/projected/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-kube-api-access-cpqtf\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128401 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128414 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d56ad5c-f3dc-4d36-862b-a68529cffacb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128427 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128438 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.128450 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4pdb\" (UniqueName: \"kubernetes.io/projected/8d56ad5c-f3dc-4d36-862b-a68529cffacb-kube-api-access-k4pdb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.303291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dd94-account-create-update-7zjm6" event={"ID":"8d56ad5c-f3dc-4d36-862b-a68529cffacb","Type":"ContainerDied","Data":"ee1203cd60e8e219f95c6e490eded27ff81b9ea7f7ba1c49a3241894e43bb25a"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.303463 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee1203cd60e8e219f95c6e490eded27ff81b9ea7f7ba1c49a3241894e43bb25a" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.303308 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dd94-account-create-update-7zjm6" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.304678 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c40d-account-create-update-snvdn" event={"ID":"229de831-b61a-4cee-8005-5a53eede49a1","Type":"ContainerDied","Data":"66a0e1fffc833a05c577b309450f6aa7c4983569fb0702a6fa029cfc32471656"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.304738 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66a0e1fffc833a05c577b309450f6aa7c4983569fb0702a6fa029cfc32471656" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.304796 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c40d-account-create-update-snvdn" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.316200 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"968fdac90b1f0897eabe537c3437ffa27f59bced52623c657229e47200b7d8e0"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.316248 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"853af3f8967f0be462a0dc793b6e4b184fdcf1d2c18300b1f013acc7ab6833c6"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.316262 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"c58974e7c9e73fc215abd7a13ee104639b223b369abc3d36ae06d348e895eb7a"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.319133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-j4rs9" event={"ID":"759a51cc-5fd3-4c7d-bfd8-072222f1e2e5","Type":"ContainerDied","Data":"aa6e27dfc901aaeab8ca6623f3d44b3ed62da84b5f7df9d6c802e3f4e63de004"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.319198 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa6e27dfc901aaeab8ca6623f3d44b3ed62da84b5f7df9d6c802e3f4e63de004" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.319384 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-j4rs9" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.321564 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-fkk8f" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.321724 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-fkk8f" event={"ID":"ac40cdd2-dadd-43ee-bf2f-d6df962d5d52","Type":"ContainerDied","Data":"d015db5968dae648e3070b7199cfb913b64a8670ccd52f0d861feed81cf196f6"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.321747 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d015db5968dae648e3070b7199cfb913b64a8670ccd52f0d861feed81cf196f6" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.323709 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-2nsvj" event={"ID":"9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635","Type":"ContainerDied","Data":"d1e5c1cd234801e5d973f978f1d6183a03143f208ce3df9bb31a5d2a3606bd8c"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.323708 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-2nsvj" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.323799 4775 scope.go:117] "RemoveContainer" containerID="2b19802da5c8df920f540957af7224105e18fec8e688835af45a27438eb0b296" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.327782 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7kcv8" event={"ID":"1ea8f43d-f87b-4f26-9d09-465c46a9e853","Type":"ContainerStarted","Data":"4bba4b9188b86331293b68e54a72fe7c36d900aa577e884627b2e9888a7e19ec"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.332360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-w8tnl" event={"ID":"a7e342a0-cce3-4690-93ef-f1afd6f8e44a","Type":"ContainerDied","Data":"c56b36128f537e202d4247f8584d0acc8df3d38417fc0ce3676faf0ac3e20cfe"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.332420 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c56b36128f537e202d4247f8584d0acc8df3d38417fc0ce3676faf0ac3e20cfe" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.332474 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-w8tnl" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.358205 4775 scope.go:117] "RemoveContainer" containerID="d3dd7e507ef69cea92870653a5ac20ba61f0f681bb9a2e2c3de82db6682d3a24" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.359113 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-7kcv8" podStartSLOduration=1.9966294310000001 podStartE2EDuration="8.359097418s" podCreationTimestamp="2025-11-26 06:41:06 +0000 UTC" firstStartedPulling="2025-11-26 06:41:07.131040425 +0000 UTC m=+950.492344367" lastFinishedPulling="2025-11-26 06:41:13.493508402 +0000 UTC m=+956.854812354" observedRunningTime="2025-11-26 06:41:14.358628945 +0000 UTC m=+957.719932917" watchObservedRunningTime="2025-11-26 06:41:14.359097418 +0000 UTC m=+957.720401360" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.365497 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c371-account-create-update-jt4mk" event={"ID":"9c651662-819c-4691-9632-5948a2b3d88c","Type":"ContainerDied","Data":"14361cf30e37e2df7a982c2c2b49e6e77d7737c66befe5d1931e879ac071ad85"} Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.365550 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14361cf30e37e2df7a982c2c2b49e6e77d7737c66befe5d1931e879ac071ad85" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.365554 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c371-account-create-update-jt4mk" Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.469994 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2nsvj"] Nov 26 06:41:14 crc kubenswrapper[4775]: I1126 06:41:14.486916 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-2nsvj"] Nov 26 06:41:15 crc kubenswrapper[4775]: I1126 06:41:15.340122 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" path="/var/lib/kubelet/pods/9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635/volumes" Nov 26 06:41:15 crc kubenswrapper[4775]: I1126 06:41:15.390482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"da76041a417097e80d68c532b9216645b1007cc84f65b2c1bed7fdc31288022c"} Nov 26 06:41:16 crc kubenswrapper[4775]: I1126 06:41:16.431169 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"0afcd81cb9935507e9d973b3416d475ad02b35db71c4f7ed8949dcff38a506b4"} Nov 26 06:41:16 crc kubenswrapper[4775]: I1126 06:41:16.431217 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"576dd63e15d3621cbea4782168add72c2c6341ff9791b375a9ea5d8937177487"} Nov 26 06:41:16 crc kubenswrapper[4775]: I1126 06:41:16.431231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"6a6554fc68758b78c8844b765e0ce6e23042f9dd34d66afde20ce8569ea3dcbe"} Nov 26 06:41:16 crc kubenswrapper[4775]: I1126 06:41:16.431241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"c244e899de1e88d96c0531a488ca8293bcda1696aed9d87745332ab8000cbbf3"} Nov 26 06:41:16 crc kubenswrapper[4775]: I1126 06:41:16.431253 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"ab1f268df7e622d3dbb459f8582dd4dccd82ef40ac05b5ca7b5fc444da6d17a9"} Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.462609 4775 generic.go:334] "Generic (PLEG): container finished" podID="1ea8f43d-f87b-4f26-9d09-465c46a9e853" containerID="4bba4b9188b86331293b68e54a72fe7c36d900aa577e884627b2e9888a7e19ec" exitCode=0 Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.462682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7kcv8" event={"ID":"1ea8f43d-f87b-4f26-9d09-465c46a9e853","Type":"ContainerDied","Data":"4bba4b9188b86331293b68e54a72fe7c36d900aa577e884627b2e9888a7e19ec"} Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.477902 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"668bb81f734971667101fb997ab361a2d92ebd42b8db9fac02192d61aae8faf8"} Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.478184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"38df1b80-132c-467f-969c-d3798dd1f7d9","Type":"ContainerStarted","Data":"8b551a16ec78a431b85a2c8a9cb61b52ae6de75781d3bd971be946031113f233"} Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.524771 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.046449348 podStartE2EDuration="48.524749362s" podCreationTimestamp="2025-11-26 06:40:29 +0000 UTC" firstStartedPulling="2025-11-26 06:41:02.866565076 +0000 UTC m=+946.227869038" lastFinishedPulling="2025-11-26 06:41:15.3448651 +0000 UTC m=+958.706169052" observedRunningTime="2025-11-26 06:41:17.515926824 +0000 UTC m=+960.877230856" watchObservedRunningTime="2025-11-26 06:41:17.524749362 +0000 UTC m=+960.886053334" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811241 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-87ltn"] Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811622 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c651662-819c-4691-9632-5948a2b3d88c" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811646 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c651662-819c-4691-9632-5948a2b3d88c" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811663 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerName="dnsmasq-dns" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811672 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerName="dnsmasq-dns" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811692 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e342a0-cce3-4690-93ef-f1afd6f8e44a" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811700 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e342a0-cce3-4690-93ef-f1afd6f8e44a" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811736 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229de831-b61a-4cee-8005-5a53eede49a1" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811745 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="229de831-b61a-4cee-8005-5a53eede49a1" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811765 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d56ad5c-f3dc-4d36-862b-a68529cffacb" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811773 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d56ad5c-f3dc-4d36-862b-a68529cffacb" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811802 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811810 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811827 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811834 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: E1126 06:41:17.811848 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerName="init" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.811855 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerName="init" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812036 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="229de831-b61a-4cee-8005-5a53eede49a1" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812062 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812077 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e342a0-cce3-4690-93ef-f1afd6f8e44a" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812093 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" containerName="mariadb-database-create" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812108 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5e0f59-7e39-43c9-b9d4-fe8dd69ec635" containerName="dnsmasq-dns" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812122 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c651662-819c-4691-9632-5948a2b3d88c" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.812134 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d56ad5c-f3dc-4d36-862b-a68529cffacb" containerName="mariadb-account-create-update" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.813146 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.815153 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.840391 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-87ltn"] Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.888764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-config\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.888821 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmkn7\" (UniqueName: \"kubernetes.io/projected/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-kube-api-access-jmkn7\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.888851 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.888873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.888893 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.888993 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.990765 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmkn7\" (UniqueName: \"kubernetes.io/projected/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-kube-api-access-jmkn7\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.990852 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.990890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.990930 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.991073 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.991135 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-config\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.994055 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.994442 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-config\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.995206 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.995427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:17 crc kubenswrapper[4775]: I1126 06:41:17.995561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.024601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmkn7\" (UniqueName: \"kubernetes.io/projected/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-kube-api-access-jmkn7\") pod \"dnsmasq-dns-7ff5475cc9-87ltn\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.137103 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.659851 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-87ltn"] Nov 26 06:41:18 crc kubenswrapper[4775]: W1126 06:41:18.674501 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod872bd7c1_51f5_4723_a5e6_c9dbbfb9b370.slice/crio-bac217fae8987605a6e3db509fd46ce299d4b25574111861deb9961c2b57f266 WatchSource:0}: Error finding container bac217fae8987605a6e3db509fd46ce299d4b25574111861deb9961c2b57f266: Status 404 returned error can't find the container with id bac217fae8987605a6e3db509fd46ce299d4b25574111861deb9961c2b57f266 Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.797041 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.907401 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-config-data\") pod \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.907493 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqtzf\" (UniqueName: \"kubernetes.io/projected/1ea8f43d-f87b-4f26-9d09-465c46a9e853-kube-api-access-xqtzf\") pod \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.907556 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-combined-ca-bundle\") pod \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\" (UID: \"1ea8f43d-f87b-4f26-9d09-465c46a9e853\") " Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.911048 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea8f43d-f87b-4f26-9d09-465c46a9e853-kube-api-access-xqtzf" (OuterVolumeSpecName: "kube-api-access-xqtzf") pod "1ea8f43d-f87b-4f26-9d09-465c46a9e853" (UID: "1ea8f43d-f87b-4f26-9d09-465c46a9e853"). InnerVolumeSpecName "kube-api-access-xqtzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.928339 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ea8f43d-f87b-4f26-9d09-465c46a9e853" (UID: "1ea8f43d-f87b-4f26-9d09-465c46a9e853"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:18 crc kubenswrapper[4775]: I1126 06:41:18.967900 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-config-data" (OuterVolumeSpecName: "config-data") pod "1ea8f43d-f87b-4f26-9d09-465c46a9e853" (UID: "1ea8f43d-f87b-4f26-9d09-465c46a9e853"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.009959 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.010000 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqtzf\" (UniqueName: \"kubernetes.io/projected/1ea8f43d-f87b-4f26-9d09-465c46a9e853-kube-api-access-xqtzf\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.010018 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ea8f43d-f87b-4f26-9d09-465c46a9e853-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.504693 4775 generic.go:334] "Generic (PLEG): container finished" podID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerID="35f8f7e6a1aa57dc6652a085cdb226ee8b57841358284cc8cd64a0b7714a6eec" exitCode=0 Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.504854 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" event={"ID":"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370","Type":"ContainerDied","Data":"35f8f7e6a1aa57dc6652a085cdb226ee8b57841358284cc8cd64a0b7714a6eec"} Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.504910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" event={"ID":"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370","Type":"ContainerStarted","Data":"bac217fae8987605a6e3db509fd46ce299d4b25574111861deb9961c2b57f266"} Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.509476 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-7kcv8" event={"ID":"1ea8f43d-f87b-4f26-9d09-465c46a9e853","Type":"ContainerDied","Data":"03ebae7f67c5f92384fbccdda6dec0b8d607fbf2052e01357db700c5828cc0f3"} Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.509510 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03ebae7f67c5f92384fbccdda6dec0b8d607fbf2052e01357db700c5828cc0f3" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.509534 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-7kcv8" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.753915 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-87ltn"] Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.779573 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tjjff"] Nov 26 06:41:19 crc kubenswrapper[4775]: E1126 06:41:19.779934 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea8f43d-f87b-4f26-9d09-465c46a9e853" containerName="keystone-db-sync" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.779956 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea8f43d-f87b-4f26-9d09-465c46a9e853" containerName="keystone-db-sync" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.780128 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea8f43d-f87b-4f26-9d09-465c46a9e853" containerName="keystone-db-sync" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.780687 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.784099 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.791403 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.791448 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.791747 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jvmv4" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.791907 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.792519 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tjjff"] Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.812114 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7"] Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.873405 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.926912 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.927179 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.927591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55dph\" (UniqueName: \"kubernetes.io/projected/faff291d-8baa-468a-8481-dbdc739cc026-kube-api-access-55dph\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.927757 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-scripts\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.927986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-fernet-keys\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.928111 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-config-data\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.928259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-config\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.928392 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvfd8\" (UniqueName: \"kubernetes.io/projected/d97295cb-cfff-4626-b1d0-0935fb711fe0-kube-api-access-nvfd8\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.928539 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-credential-keys\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.928933 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.929023 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.929106 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-combined-ca-bundle\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.931821 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7"] Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.996995 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-ng8zs"] Nov 26 06:41:19 crc kubenswrapper[4775]: I1126 06:41:19.998039 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.002443 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dkl9z" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.002636 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.002775 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.004631 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-mqh7n"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.005765 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.008443 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zq8z7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.009552 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.009699 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.014172 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ng8zs"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.025435 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mqh7n"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-scripts\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030492 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-fernet-keys\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030515 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-config-data\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-config\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030565 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvfd8\" (UniqueName: \"kubernetes.io/projected/d97295cb-cfff-4626-b1d0-0935fb711fe0-kube-api-access-nvfd8\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-credential-keys\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030615 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030647 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030664 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-combined-ca-bundle\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030758 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.030790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55dph\" (UniqueName: \"kubernetes.io/projected/faff291d-8baa-468a-8481-dbdc739cc026-kube-api-access-55dph\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.042189 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.042866 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.042922 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f8db555df-xs5d8"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.043158 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-config\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.043505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.044090 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.044675 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.049807 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f8db555df-xs5d8"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.049847 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.050444 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.050578 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.054047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-fernet-keys\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.054428 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-lh7zn" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.055449 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-combined-ca-bundle\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.056324 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-credential-keys\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.057135 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-scripts\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.057799 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-config-data\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.089980 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55dph\" (UniqueName: \"kubernetes.io/projected/faff291d-8baa-468a-8481-dbdc739cc026-kube-api-access-55dph\") pod \"dnsmasq-dns-5c5cc7c5ff-7pzw7\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.110278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvfd8\" (UniqueName: \"kubernetes.io/projected/d97295cb-cfff-4626-b1d0-0935fb711fe0-kube-api-access-nvfd8\") pod \"keystone-bootstrap-tjjff\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.118500 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.133885 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-scripts\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.133942 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-combined-ca-bundle\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.133986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-scripts\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134017 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-horizon-secret-key\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134064 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-config-data\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134083 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-db-sync-config-data\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134122 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-config-data\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134150 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-config\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134171 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-combined-ca-bundle\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvt6r\" (UniqueName: \"kubernetes.io/projected/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-kube-api-access-gvt6r\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134239 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nntvk\" (UniqueName: \"kubernetes.io/projected/747654d0-4f51-411e-958b-34f3cd5403eb-kube-api-access-nntvk\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134289 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzfbs\" (UniqueName: \"kubernetes.io/projected/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-kube-api-access-tzfbs\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134347 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-logs\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.134371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-etc-machine-id\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.179560 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-f8988"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.181203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.185221 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d888c" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.191785 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.194307 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f8988"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.228739 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-n9tbr"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.229114 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-combined-ca-bundle\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235797 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-scripts\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235822 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-combined-ca-bundle\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235847 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-horizon-secret-key\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235881 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-config-data\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-db-sync-config-data\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235925 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-config-data\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235943 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7765k\" (UniqueName: \"kubernetes.io/projected/25784e47-2e5a-4df0-9aa2-c2a42659cb40-kube-api-access-7765k\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.235987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-config\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-combined-ca-bundle\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236026 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvt6r\" (UniqueName: \"kubernetes.io/projected/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-kube-api-access-gvt6r\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236057 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nntvk\" (UniqueName: \"kubernetes.io/projected/747654d0-4f51-411e-958b-34f3cd5403eb-kube-api-access-nntvk\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236090 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-db-sync-config-data\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236107 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzfbs\" (UniqueName: \"kubernetes.io/projected/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-kube-api-access-tzfbs\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-logs\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-etc-machine-id\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236184 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-scripts\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.236896 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-scripts\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.238984 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-logs\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.238790 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-etc-machine-id\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.241541 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.241747 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-config-data\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.243945 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-combined-ca-bundle\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.246536 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.246697 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.246948 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zmxnc" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.253733 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7645c8b599-r5tdh"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.255270 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.256449 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-config\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.257850 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-horizon-secret-key\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.258248 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-db-sync-config-data\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.267773 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-combined-ca-bundle\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.268249 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-scripts\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.268318 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-n9tbr"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.268834 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7645c8b599-r5tdh"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.273632 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-config-data\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.281594 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.287404 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzfbs\" (UniqueName: \"kubernetes.io/projected/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-kube-api-access-tzfbs\") pod \"horizon-f8db555df-xs5d8\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.306700 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvt6r\" (UniqueName: \"kubernetes.io/projected/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-kube-api-access-gvt6r\") pod \"cinder-db-sync-ng8zs\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.307147 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nntvk\" (UniqueName: \"kubernetes.io/projected/747654d0-4f51-411e-958b-34f3cd5403eb-kube-api-access-nntvk\") pod \"neutron-db-sync-mqh7n\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.323757 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.325619 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.333214 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.334042 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342233 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f163bf-e900-4d6e-b5be-948f927c7b86-logs\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-combined-ca-bundle\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-config-data\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7765k\" (UniqueName: \"kubernetes.io/projected/25784e47-2e5a-4df0-9aa2-c2a42659cb40-kube-api-access-7765k\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342422 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-config-data\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23563003-3975-4d09-aec4-3b3004aa3672-horizon-secret-key\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342928 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-scripts\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.342975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-db-sync-config-data\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.343257 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2prm6\" (UniqueName: \"kubernetes.io/projected/23563003-3975-4d09-aec4-3b3004aa3672-kube-api-access-2prm6\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.343390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-combined-ca-bundle\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.343832 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5h54\" (UniqueName: \"kubernetes.io/projected/b0f163bf-e900-4d6e-b5be-948f927c7b86-kube-api-access-b5h54\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.343866 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-scripts\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.343896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23563003-3975-4d09-aec4-3b3004aa3672-logs\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.350747 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.355896 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.361385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-db-sync-config-data\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.361931 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-combined-ca-bundle\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.362598 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-cs4rl"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.363924 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.377677 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-cs4rl"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.380506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7765k\" (UniqueName: \"kubernetes.io/projected/25784e47-2e5a-4df0-9aa2-c2a42659cb40-kube-api-access-7765k\") pod \"barbican-db-sync-f8988\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.393940 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.395617 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.400551 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.400980 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qtdst" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.401210 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.402020 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8988" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.405763 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.411105 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-scripts\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451772 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451810 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f163bf-e900-4d6e-b5be-948f927c7b86-logs\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451859 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-config-data\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5wwk\" (UniqueName: \"kubernetes.io/projected/a14e76fa-5315-4564-9430-9231a8e31a01-kube-api-access-g5wwk\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451900 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451943 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-logs\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.451986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msnr2\" (UniqueName: \"kubernetes.io/projected/acbdc4be-0331-410a-8f6c-b87323d59456-kube-api-access-msnr2\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452046 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-config-data\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452084 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23563003-3975-4d09-aec4-3b3004aa3672-horizon-secret-key\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452147 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-scripts\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452202 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-config\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-log-httpd\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452246 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-config-data\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452282 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2prm6\" (UniqueName: \"kubernetes.io/projected/23563003-3975-4d09-aec4-3b3004aa3672-kube-api-access-2prm6\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-combined-ca-bundle\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f163bf-e900-4d6e-b5be-948f927c7b86-logs\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452318 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452359 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5h54\" (UniqueName: \"kubernetes.io/projected/b0f163bf-e900-4d6e-b5be-948f927c7b86-kube-api-access-b5h54\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452386 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-scripts\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452401 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm9cd\" (UniqueName: \"kubernetes.io/projected/1d277cfa-56f7-4692-8724-980d3f6dbe0c-kube-api-access-lm9cd\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452433 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23563003-3975-4d09-aec4-3b3004aa3672-logs\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.452453 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-run-httpd\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.454146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-scripts\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.454678 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.457650 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.459569 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-config-data\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.460867 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.461218 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.465449 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23563003-3975-4d09-aec4-3b3004aa3672-logs\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.468076 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.468351 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-combined-ca-bundle\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.472731 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-scripts\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.474777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23563003-3975-4d09-aec4-3b3004aa3672-horizon-secret-key\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.475786 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2prm6\" (UniqueName: \"kubernetes.io/projected/23563003-3975-4d09-aec4-3b3004aa3672-kube-api-access-2prm6\") pod \"horizon-7645c8b599-r5tdh\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.492791 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.493450 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-config-data\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.496859 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.497851 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5h54\" (UniqueName: \"kubernetes.io/projected/b0f163bf-e900-4d6e-b5be-948f927c7b86-kube-api-access-b5h54\") pod \"placement-db-sync-n9tbr\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.525695 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n9tbr" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.548259 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.556192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" event={"ID":"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370","Type":"ContainerStarted","Data":"f3fe7fa2297315daa9c684403116b63845fccd321c47ef32a51c95d2c5b70134"} Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.556374 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerName="dnsmasq-dns" containerID="cri-o://f3fe7fa2297315daa9c684403116b63845fccd321c47ef32a51c95d2c5b70134" gracePeriod=10 Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.556652 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557093 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557133 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557203 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557263 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557283 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-config\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-log-httpd\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-config-data\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557416 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557454 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm9cd\" (UniqueName: \"kubernetes.io/projected/1d277cfa-56f7-4692-8724-980d3f6dbe0c-kube-api-access-lm9cd\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557519 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79m8j\" (UniqueName: \"kubernetes.io/projected/31d04adf-82d4-4643-97df-8f6d9672b4cb-kube-api-access-79m8j\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557542 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557561 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-run-httpd\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-scripts\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557611 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557647 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557671 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5wwk\" (UniqueName: \"kubernetes.io/projected/a14e76fa-5315-4564-9430-9231a8e31a01-kube-api-access-g5wwk\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557708 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557761 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-logs\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-logs\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557805 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557819 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.557839 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msnr2\" (UniqueName: \"kubernetes.io/projected/acbdc4be-0331-410a-8f6c-b87323d59456-kube-api-access-msnr2\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.558892 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.559281 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-run-httpd\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.559733 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.560533 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-config\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.561011 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-log-httpd\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.562851 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.570071 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.570548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-logs\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.570988 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-scripts\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.571524 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-config-data\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.577700 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.593664 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.598000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.598122 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-config-data\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.598275 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.598796 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-scripts\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.599515 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msnr2\" (UniqueName: \"kubernetes.io/projected/acbdc4be-0331-410a-8f6c-b87323d59456-kube-api-access-msnr2\") pod \"dnsmasq-dns-8b5c85b87-cs4rl\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.600872 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.605165 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.605256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.607489 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm9cd\" (UniqueName: \"kubernetes.io/projected/1d277cfa-56f7-4692-8724-980d3f6dbe0c-kube-api-access-lm9cd\") pod \"glance-default-external-api-0\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.627158 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5wwk\" (UniqueName: \"kubernetes.io/projected/a14e76fa-5315-4564-9430-9231a8e31a01-kube-api-access-g5wwk\") pod \"ceilometer-0\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.669370 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.675046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.675205 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79m8j\" (UniqueName: \"kubernetes.io/projected/31d04adf-82d4-4643-97df-8f6d9672b4cb-kube-api-access-79m8j\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.675292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.675441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.675938 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-logs\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.676069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.676199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.676356 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-logs\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.676089 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.682563 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.684515 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.684744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.687398 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.688132 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.688482 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.695863 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79m8j\" (UniqueName: \"kubernetes.io/projected/31d04adf-82d4-4643-97df-8f6d9672b4cb-kube-api-access-79m8j\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.713657 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.754128 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.858639 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.929977 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" podStartSLOduration=3.929958896 podStartE2EDuration="3.929958896s" podCreationTimestamp="2025-11-26 06:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:20.597885935 +0000 UTC m=+963.959189897" watchObservedRunningTime="2025-11-26 06:41:20.929958896 +0000 UTC m=+964.291262848" Nov 26 06:41:20 crc kubenswrapper[4775]: I1126 06:41:20.998968 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7"] Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.049062 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.480585 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tjjff"] Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.526083 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7645c8b599-r5tdh"] Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.545348 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-ng8zs"] Nov 26 06:41:21 crc kubenswrapper[4775]: W1126 06:41:21.572801 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d6b7cfb_7b59_4f0b_b36c_94dae19e512d.slice/crio-167037b719d8993e64892ba22bab0fb6af36e113f3882870ea0f70d0027d3154 WatchSource:0}: Error finding container 167037b719d8993e64892ba22bab0fb6af36e113f3882870ea0f70d0027d3154: Status 404 returned error can't find the container with id 167037b719d8993e64892ba22bab0fb6af36e113f3882870ea0f70d0027d3154 Nov 26 06:41:21 crc kubenswrapper[4775]: W1126 06:41:21.585204 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25784e47_2e5a_4df0_9aa2_c2a42659cb40.slice/crio-63fad62ec314c274ede68d1ca00678d6c452d6fec1f4bc728e53dc9b5b0b2548 WatchSource:0}: Error finding container 63fad62ec314c274ede68d1ca00678d6c452d6fec1f4bc728e53dc9b5b0b2548: Status 404 returned error can't find the container with id 63fad62ec314c274ede68d1ca00678d6c452d6fec1f4bc728e53dc9b5b0b2548 Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.585780 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f8988"] Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.605642 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" event={"ID":"faff291d-8baa-468a-8481-dbdc739cc026","Type":"ContainerStarted","Data":"8b4b99c9341ddc94b99567a10581e82b48009476d26bbf45d14a9d287ac23535"} Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.605680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" event={"ID":"faff291d-8baa-468a-8481-dbdc739cc026","Type":"ContainerStarted","Data":"5f6dc8b328b3a8e2e7fbe782ebde93bc442b5be72e3664aabfbf803fcd5b9cba"} Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.607229 4775 generic.go:334] "Generic (PLEG): container finished" podID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerID="f3fe7fa2297315daa9c684403116b63845fccd321c47ef32a51c95d2c5b70134" exitCode=0 Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.607271 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" event={"ID":"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370","Type":"ContainerDied","Data":"f3fe7fa2297315daa9c684403116b63845fccd321c47ef32a51c95d2c5b70134"} Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.607966 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tjjff" event={"ID":"d97295cb-cfff-4626-b1d0-0935fb711fe0","Type":"ContainerStarted","Data":"9e972185132634907c02a609cc60744aded18ff02b72ea698730d58b83c5cd25"} Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.614264 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7645c8b599-r5tdh" event={"ID":"23563003-3975-4d09-aec4-3b3004aa3672","Type":"ContainerStarted","Data":"1be8fbe7e407e66dbf62cebfeafe4a4eba89fc648a8e18f6e7b24752a763f6e8"} Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.680763 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.710831 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-swift-storage-0\") pod \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.710897 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-config\") pod \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.710973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-svc\") pod \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.711006 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-nb\") pod \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.711046 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-sb\") pod \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.711096 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmkn7\" (UniqueName: \"kubernetes.io/projected/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-kube-api-access-jmkn7\") pod \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\" (UID: \"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370\") " Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.723010 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f8db555df-xs5d8"] Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.748161 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-kube-api-access-jmkn7" (OuterVolumeSpecName: "kube-api-access-jmkn7") pod "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" (UID: "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370"). InnerVolumeSpecName "kube-api-access-jmkn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.766013 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-mqh7n"] Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.815972 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmkn7\" (UniqueName: \"kubernetes.io/projected/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-kube-api-access-jmkn7\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.834957 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-config" (OuterVolumeSpecName: "config") pod "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" (UID: "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.835707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" (UID: "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.844673 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" (UID: "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.847178 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" (UID: "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.849111 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" (UID: "872bd7c1-51f5-4723-a5e6-c9dbbfb9b370"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.918351 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.918409 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.918427 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.918447 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:21 crc kubenswrapper[4775]: I1126 06:41:21.918467 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.016048 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-n9tbr"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.029187 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.041733 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-cs4rl"] Nov 26 06:41:22 crc kubenswrapper[4775]: W1126 06:41:22.091191 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0f163bf_e900_4d6e_b5be_948f927c7b86.slice/crio-d074a1f62bff62a1204c7ed11717deeeb1eb68207a9d489917399041e191b0e3 WatchSource:0}: Error finding container d074a1f62bff62a1204c7ed11717deeeb1eb68207a9d489917399041e191b0e3: Status 404 returned error can't find the container with id d074a1f62bff62a1204c7ed11717deeeb1eb68207a9d489917399041e191b0e3 Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.214399 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.257791 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.264982 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f8db555df-xs5d8"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.296191 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.319226 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:22 crc kubenswrapper[4775]: W1126 06:41:22.336161 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d277cfa_56f7_4692_8724_980d3f6dbe0c.slice/crio-960ed5f5772ecf1902800838043f952847cd5cef84e9bed7362fa81d18829f1b WatchSource:0}: Error finding container 960ed5f5772ecf1902800838043f952847cd5cef84e9bed7362fa81d18829f1b: Status 404 returned error can't find the container with id 960ed5f5772ecf1902800838043f952847cd5cef84e9bed7362fa81d18829f1b Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.368369 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-84989d94bf-7vmkz"] Nov 26 06:41:22 crc kubenswrapper[4775]: E1126 06:41:22.368933 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerName="init" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.368954 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerName="init" Nov 26 06:41:22 crc kubenswrapper[4775]: E1126 06:41:22.368988 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerName="dnsmasq-dns" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.368996 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerName="dnsmasq-dns" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.369243 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" containerName="dnsmasq-dns" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.371196 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.398455 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84989d94bf-7vmkz"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.431850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rfxs\" (UniqueName: \"kubernetes.io/projected/10c96169-294e-4273-8e08-92c677dfc01c-kube-api-access-6rfxs\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.432007 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-config-data\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.432062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/10c96169-294e-4273-8e08-92c677dfc01c-horizon-secret-key\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.432131 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.432155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c96169-294e-4273-8e08-92c677dfc01c-logs\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.538915 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-config-data\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.538997 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/10c96169-294e-4273-8e08-92c677dfc01c-horizon-secret-key\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.539070 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.539097 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c96169-294e-4273-8e08-92c677dfc01c-logs\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.539140 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rfxs\" (UniqueName: \"kubernetes.io/projected/10c96169-294e-4273-8e08-92c677dfc01c-kube-api-access-6rfxs\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.541161 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c96169-294e-4273-8e08-92c677dfc01c-logs\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.541421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.545548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-config-data\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.556360 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/10c96169-294e-4273-8e08-92c677dfc01c-horizon-secret-key\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.567979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rfxs\" (UniqueName: \"kubernetes.io/projected/10c96169-294e-4273-8e08-92c677dfc01c-kube-api-access-6rfxs\") pod \"horizon-84989d94bf-7vmkz\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.634960 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n9tbr" event={"ID":"b0f163bf-e900-4d6e-b5be-948f927c7b86","Type":"ContainerStarted","Data":"d074a1f62bff62a1204c7ed11717deeeb1eb68207a9d489917399041e191b0e3"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.636555 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" event={"ID":"acbdc4be-0331-410a-8f6c-b87323d59456","Type":"ContainerStarted","Data":"04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.636591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" event={"ID":"acbdc4be-0331-410a-8f6c-b87323d59456","Type":"ContainerStarted","Data":"1240e7f8f1ab4980af1572bd8249739fa485ff91505dd51633cae09c4e743bf1"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.641962 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tjjff" event={"ID":"d97295cb-cfff-4626-b1d0-0935fb711fe0","Type":"ContainerStarted","Data":"f8bcd7ec20eb8a4ae97d80fe889b3e59b2b484ea35b4fbc229a8c5ede4cf3cfb"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.644856 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8988" event={"ID":"25784e47-2e5a-4df0-9aa2-c2a42659cb40","Type":"ContainerStarted","Data":"63fad62ec314c274ede68d1ca00678d6c452d6fec1f4bc728e53dc9b5b0b2548"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.646770 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerStarted","Data":"966d923c6267562e6b22c451aa680e25509daf2a420c16f716f81e13ba366f8f"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.670035 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqh7n" event={"ID":"747654d0-4f51-411e-958b-34f3cd5403eb","Type":"ContainerStarted","Data":"b13fda92c3c8312d7063282c76057773adc40ebd3ea6936dc6aebb2f25e040ec"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.670079 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqh7n" event={"ID":"747654d0-4f51-411e-958b-34f3cd5403eb","Type":"ContainerStarted","Data":"07b4e70c36ff7fcff560dc391cb74e581def5583117ac13a90adbef3f460cac1"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.673643 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d277cfa-56f7-4692-8724-980d3f6dbe0c","Type":"ContainerStarted","Data":"960ed5f5772ecf1902800838043f952847cd5cef84e9bed7362fa81d18829f1b"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.700861 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.701408 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-87ltn" event={"ID":"872bd7c1-51f5-4723-a5e6-c9dbbfb9b370","Type":"ContainerDied","Data":"bac217fae8987605a6e3db509fd46ce299d4b25574111861deb9961c2b57f266"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.701481 4775 scope.go:117] "RemoveContainer" containerID="f3fe7fa2297315daa9c684403116b63845fccd321c47ef32a51c95d2c5b70134" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.709173 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tjjff" podStartSLOduration=3.70915313 podStartE2EDuration="3.70915313s" podCreationTimestamp="2025-11-26 06:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:22.706702754 +0000 UTC m=+966.068006706" watchObservedRunningTime="2025-11-26 06:41:22.70915313 +0000 UTC m=+966.070457082" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.710121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ng8zs" event={"ID":"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d","Type":"ContainerStarted","Data":"167037b719d8993e64892ba22bab0fb6af36e113f3882870ea0f70d0027d3154"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.719093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8db555df-xs5d8" event={"ID":"b0a4cb55-482a-4be2-9c7b-48874b6cbbac","Type":"ContainerStarted","Data":"39ef0677f1dda6eeeb348bc0da9095c1a903d8555c21722e810d2296a7169b50"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.721313 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.729579 4775 generic.go:334] "Generic (PLEG): container finished" podID="faff291d-8baa-468a-8481-dbdc739cc026" containerID="8b4b99c9341ddc94b99567a10581e82b48009476d26bbf45d14a9d287ac23535" exitCode=0 Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.729995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" event={"ID":"faff291d-8baa-468a-8481-dbdc739cc026","Type":"ContainerDied","Data":"8b4b99c9341ddc94b99567a10581e82b48009476d26bbf45d14a9d287ac23535"} Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.744866 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-mqh7n" podStartSLOduration=3.74478107 podStartE2EDuration="3.74478107s" podCreationTimestamp="2025-11-26 06:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:22.723565418 +0000 UTC m=+966.084869390" watchObservedRunningTime="2025-11-26 06:41:22.74478107 +0000 UTC m=+966.106085032" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.759165 4775 scope.go:117] "RemoveContainer" containerID="35f8f7e6a1aa57dc6652a085cdb226ee8b57841358284cc8cd64a0b7714a6eec" Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.780243 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-87ltn"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.795541 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-87ltn"] Nov 26 06:41:22 crc kubenswrapper[4775]: I1126 06:41:22.913510 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.234781 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.348553 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="872bd7c1-51f5-4723-a5e6-c9dbbfb9b370" path="/var/lib/kubelet/pods/872bd7c1-51f5-4723-a5e6-c9dbbfb9b370/volumes" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.374814 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-config\") pod \"faff291d-8baa-468a-8481-dbdc739cc026\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.374940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-svc\") pod \"faff291d-8baa-468a-8481-dbdc739cc026\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.374983 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-nb\") pod \"faff291d-8baa-468a-8481-dbdc739cc026\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.375038 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-sb\") pod \"faff291d-8baa-468a-8481-dbdc739cc026\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.375066 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55dph\" (UniqueName: \"kubernetes.io/projected/faff291d-8baa-468a-8481-dbdc739cc026-kube-api-access-55dph\") pod \"faff291d-8baa-468a-8481-dbdc739cc026\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.375167 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-swift-storage-0\") pod \"faff291d-8baa-468a-8481-dbdc739cc026\" (UID: \"faff291d-8baa-468a-8481-dbdc739cc026\") " Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.407702 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faff291d-8baa-468a-8481-dbdc739cc026-kube-api-access-55dph" (OuterVolumeSpecName: "kube-api-access-55dph") pod "faff291d-8baa-468a-8481-dbdc739cc026" (UID: "faff291d-8baa-468a-8481-dbdc739cc026"). InnerVolumeSpecName "kube-api-access-55dph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.409316 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "faff291d-8baa-468a-8481-dbdc739cc026" (UID: "faff291d-8baa-468a-8481-dbdc739cc026"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.415024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-config" (OuterVolumeSpecName: "config") pod "faff291d-8baa-468a-8481-dbdc739cc026" (UID: "faff291d-8baa-468a-8481-dbdc739cc026"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.420496 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84989d94bf-7vmkz"] Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.438806 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "faff291d-8baa-468a-8481-dbdc739cc026" (UID: "faff291d-8baa-468a-8481-dbdc739cc026"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.443984 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "faff291d-8baa-468a-8481-dbdc739cc026" (UID: "faff291d-8baa-468a-8481-dbdc739cc026"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.458496 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "faff291d-8baa-468a-8481-dbdc739cc026" (UID: "faff291d-8baa-468a-8481-dbdc739cc026"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.485877 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.485908 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.485919 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.485929 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.485940 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55dph\" (UniqueName: \"kubernetes.io/projected/faff291d-8baa-468a-8481-dbdc739cc026-kube-api-access-55dph\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.485949 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/faff291d-8baa-468a-8481-dbdc739cc026-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.746112 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d277cfa-56f7-4692-8724-980d3f6dbe0c","Type":"ContainerStarted","Data":"d5c7cdda2820e517050f9407209991a6aed7a48b7397a085a2eb82d0a47a5223"} Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.751768 4775 generic.go:334] "Generic (PLEG): container finished" podID="acbdc4be-0331-410a-8f6c-b87323d59456" containerID="04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba" exitCode=0 Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.751845 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" event={"ID":"acbdc4be-0331-410a-8f6c-b87323d59456","Type":"ContainerDied","Data":"04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba"} Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.753990 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84989d94bf-7vmkz" event={"ID":"10c96169-294e-4273-8e08-92c677dfc01c","Type":"ContainerStarted","Data":"ca53cabd1ed7c8756b01398673f8a23343b8cc29c51f966fa0f9917fd93a18dd"} Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.758451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31d04adf-82d4-4643-97df-8f6d9672b4cb","Type":"ContainerStarted","Data":"f718906d2408ae959e644438577df8098db1dffbbefdf74f1b8b10c25cb0922e"} Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.762227 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" event={"ID":"faff291d-8baa-468a-8481-dbdc739cc026","Type":"ContainerDied","Data":"5f6dc8b328b3a8e2e7fbe782ebde93bc442b5be72e3664aabfbf803fcd5b9cba"} Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.762267 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.762294 4775 scope.go:117] "RemoveContainer" containerID="8b4b99c9341ddc94b99567a10581e82b48009476d26bbf45d14a9d287ac23535" Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.884466 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7"] Nov 26 06:41:23 crc kubenswrapper[4775]: I1126 06:41:23.891234 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-7pzw7"] Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.805177 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d277cfa-56f7-4692-8724-980d3f6dbe0c","Type":"ContainerStarted","Data":"5c194b3dc440f7ca36cac1595ee1d8784f44529eb31faf058772202a181c0519"} Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.805242 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-log" containerID="cri-o://d5c7cdda2820e517050f9407209991a6aed7a48b7397a085a2eb82d0a47a5223" gracePeriod=30 Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.805321 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-httpd" containerID="cri-o://5c194b3dc440f7ca36cac1595ee1d8784f44529eb31faf058772202a181c0519" gracePeriod=30 Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.810878 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" event={"ID":"acbdc4be-0331-410a-8f6c-b87323d59456","Type":"ContainerStarted","Data":"75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2"} Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.811171 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.813293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31d04adf-82d4-4643-97df-8f6d9672b4cb","Type":"ContainerStarted","Data":"bb28b80f049fe85c5df266f2d58eb9c20ac620fb7a9b283daed1076a7459f2fc"} Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.846052 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.846014544 podStartE2EDuration="4.846014544s" podCreationTimestamp="2025-11-26 06:41:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:24.843490876 +0000 UTC m=+968.204794828" watchObservedRunningTime="2025-11-26 06:41:24.846014544 +0000 UTC m=+968.207318496" Nov 26 06:41:24 crc kubenswrapper[4775]: I1126 06:41:24.867747 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" podStartSLOduration=4.867732149 podStartE2EDuration="4.867732149s" podCreationTimestamp="2025-11-26 06:41:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:24.864661766 +0000 UTC m=+968.225965738" watchObservedRunningTime="2025-11-26 06:41:24.867732149 +0000 UTC m=+968.229036101" Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.343273 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faff291d-8baa-468a-8481-dbdc739cc026" path="/var/lib/kubelet/pods/faff291d-8baa-468a-8481-dbdc739cc026/volumes" Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.826832 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31d04adf-82d4-4643-97df-8f6d9672b4cb","Type":"ContainerStarted","Data":"8282484c093c2fc8a5678ed3717256fdb144f4b27a85f2b08f8ffb603b8df6de"} Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.826899 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-log" containerID="cri-o://bb28b80f049fe85c5df266f2d58eb9c20ac620fb7a9b283daed1076a7459f2fc" gracePeriod=30 Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.826951 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-httpd" containerID="cri-o://8282484c093c2fc8a5678ed3717256fdb144f4b27a85f2b08f8ffb603b8df6de" gracePeriod=30 Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.832701 4775 generic.go:334] "Generic (PLEG): container finished" podID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerID="5c194b3dc440f7ca36cac1595ee1d8784f44529eb31faf058772202a181c0519" exitCode=143 Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.832739 4775 generic.go:334] "Generic (PLEG): container finished" podID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerID="d5c7cdda2820e517050f9407209991a6aed7a48b7397a085a2eb82d0a47a5223" exitCode=143 Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.832738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d277cfa-56f7-4692-8724-980d3f6dbe0c","Type":"ContainerDied","Data":"5c194b3dc440f7ca36cac1595ee1d8784f44529eb31faf058772202a181c0519"} Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.832771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d277cfa-56f7-4692-8724-980d3f6dbe0c","Type":"ContainerDied","Data":"d5c7cdda2820e517050f9407209991a6aed7a48b7397a085a2eb82d0a47a5223"} Nov 26 06:41:25 crc kubenswrapper[4775]: I1126 06:41:25.852347 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.852331349 podStartE2EDuration="5.852331349s" podCreationTimestamp="2025-11-26 06:41:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:25.846653456 +0000 UTC m=+969.207957418" watchObservedRunningTime="2025-11-26 06:41:25.852331349 +0000 UTC m=+969.213635301" Nov 26 06:41:26 crc kubenswrapper[4775]: I1126 06:41:26.846109 4775 generic.go:334] "Generic (PLEG): container finished" podID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerID="8282484c093c2fc8a5678ed3717256fdb144f4b27a85f2b08f8ffb603b8df6de" exitCode=0 Nov 26 06:41:26 crc kubenswrapper[4775]: I1126 06:41:26.846415 4775 generic.go:334] "Generic (PLEG): container finished" podID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerID="bb28b80f049fe85c5df266f2d58eb9c20ac620fb7a9b283daed1076a7459f2fc" exitCode=143 Nov 26 06:41:26 crc kubenswrapper[4775]: I1126 06:41:26.846442 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31d04adf-82d4-4643-97df-8f6d9672b4cb","Type":"ContainerDied","Data":"8282484c093c2fc8a5678ed3717256fdb144f4b27a85f2b08f8ffb603b8df6de"} Nov 26 06:41:26 crc kubenswrapper[4775]: I1126 06:41:26.846472 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31d04adf-82d4-4643-97df-8f6d9672b4cb","Type":"ContainerDied","Data":"bb28b80f049fe85c5df266f2d58eb9c20ac620fb7a9b283daed1076a7459f2fc"} Nov 26 06:41:27 crc kubenswrapper[4775]: I1126 06:41:27.856042 4775 generic.go:334] "Generic (PLEG): container finished" podID="d97295cb-cfff-4626-b1d0-0935fb711fe0" containerID="f8bcd7ec20eb8a4ae97d80fe889b3e59b2b484ea35b4fbc229a8c5ede4cf3cfb" exitCode=0 Nov 26 06:41:27 crc kubenswrapper[4775]: I1126 06:41:27.856264 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tjjff" event={"ID":"d97295cb-cfff-4626-b1d0-0935fb711fe0","Type":"ContainerDied","Data":"f8bcd7ec20eb8a4ae97d80fe889b3e59b2b484ea35b4fbc229a8c5ede4cf3cfb"} Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.936005 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7645c8b599-r5tdh"] Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.960861 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5886fff4d-zcj7s"] Nov 26 06:41:28 crc kubenswrapper[4775]: E1126 06:41:28.961324 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faff291d-8baa-468a-8481-dbdc739cc026" containerName="init" Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.961350 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="faff291d-8baa-468a-8481-dbdc739cc026" containerName="init" Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.961585 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="faff291d-8baa-468a-8481-dbdc739cc026" containerName="init" Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.962515 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.965178 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.977764 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5886fff4d-zcj7s"] Nov 26 06:41:28 crc kubenswrapper[4775]: I1126 06:41:28.991816 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84989d94bf-7vmkz"] Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.047014 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-9c9599df6-zd8gh"] Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.048527 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.106242 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9c9599df6-zd8gh"] Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115327 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-tls-certs\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115382 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-combined-ca-bundle\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115481 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vltpp\" (UniqueName: \"kubernetes.io/projected/fb607340-2aa9-4b57-a4b7-a6a64457cae3-kube-api-access-vltpp\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115748 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-scripts\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115841 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-config-data\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb607340-2aa9-4b57-a4b7-a6a64457cae3-logs\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.115976 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-secret-key\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.218061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f578a3d5-f846-44b6-bc2b-3d98188dffff-scripts\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.218149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-config-data\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.219538 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-config-data\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.219586 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f578a3d5-f846-44b6-bc2b-3d98188dffff-logs\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.219660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-combined-ca-bundle\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.219757 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb607340-2aa9-4b57-a4b7-a6a64457cae3-logs\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.219928 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-horizon-secret-key\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220027 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-secret-key\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220092 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-horizon-tls-certs\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb607340-2aa9-4b57-a4b7-a6a64457cae3-logs\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220200 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-tls-certs\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-combined-ca-bundle\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f578a3d5-f846-44b6-bc2b-3d98188dffff-config-data\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220294 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw88v\" (UniqueName: \"kubernetes.io/projected/f578a3d5-f846-44b6-bc2b-3d98188dffff-kube-api-access-kw88v\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220358 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vltpp\" (UniqueName: \"kubernetes.io/projected/fb607340-2aa9-4b57-a4b7-a6a64457cae3-kube-api-access-vltpp\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.220549 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-scripts\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.221036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-scripts\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.227305 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-secret-key\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.227796 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-combined-ca-bundle\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.228390 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-tls-certs\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.234851 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vltpp\" (UniqueName: \"kubernetes.io/projected/fb607340-2aa9-4b57-a4b7-a6a64457cae3-kube-api-access-vltpp\") pod \"horizon-5886fff4d-zcj7s\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.287799 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.321961 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-horizon-tls-certs\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322048 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f578a3d5-f846-44b6-bc2b-3d98188dffff-config-data\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw88v\" (UniqueName: \"kubernetes.io/projected/f578a3d5-f846-44b6-bc2b-3d98188dffff-kube-api-access-kw88v\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322113 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f578a3d5-f846-44b6-bc2b-3d98188dffff-scripts\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322150 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f578a3d5-f846-44b6-bc2b-3d98188dffff-logs\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322176 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-combined-ca-bundle\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322216 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-horizon-secret-key\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.322949 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f578a3d5-f846-44b6-bc2b-3d98188dffff-logs\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.323379 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f578a3d5-f846-44b6-bc2b-3d98188dffff-scripts\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.324057 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f578a3d5-f846-44b6-bc2b-3d98188dffff-config-data\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.328536 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-horizon-secret-key\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.328631 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-combined-ca-bundle\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.328672 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f578a3d5-f846-44b6-bc2b-3d98188dffff-horizon-tls-certs\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.340048 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw88v\" (UniqueName: \"kubernetes.io/projected/f578a3d5-f846-44b6-bc2b-3d98188dffff-kube-api-access-kw88v\") pod \"horizon-9c9599df6-zd8gh\" (UID: \"f578a3d5-f846-44b6-bc2b-3d98188dffff\") " pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:29 crc kubenswrapper[4775]: I1126 06:41:29.378929 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:30 crc kubenswrapper[4775]: I1126 06:41:30.689889 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:41:30 crc kubenswrapper[4775]: I1126 06:41:30.765612 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jggzv"] Nov 26 06:41:30 crc kubenswrapper[4775]: I1126 06:41:30.765918 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" containerID="cri-o://e8aebe0fab1e656480dc0057b764996044e1fa9b766e37f0b917ec6ebbed7bcc" gracePeriod=10 Nov 26 06:41:31 crc kubenswrapper[4775]: I1126 06:41:31.331093 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Nov 26 06:41:31 crc kubenswrapper[4775]: I1126 06:41:31.936564 4775 generic.go:334] "Generic (PLEG): container finished" podID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerID="e8aebe0fab1e656480dc0057b764996044e1fa9b766e37f0b917ec6ebbed7bcc" exitCode=0 Nov 26 06:41:31 crc kubenswrapper[4775]: I1126 06:41:31.936678 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" event={"ID":"ed38647f-f4ed-4993-83f1-f7f34255e9b7","Type":"ContainerDied","Data":"e8aebe0fab1e656480dc0057b764996044e1fa9b766e37f0b917ec6ebbed7bcc"} Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.330924 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Nov 26 06:41:36 crc kubenswrapper[4775]: E1126 06:41:36.560815 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 26 06:41:36 crc kubenswrapper[4775]: E1126 06:41:36.561029 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7765k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-f8988_openstack(25784e47-2e5a-4df0-9aa2-c2a42659cb40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:41:36 crc kubenswrapper[4775]: E1126 06:41:36.563142 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-f8988" podUID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" Nov 26 06:41:36 crc kubenswrapper[4775]: E1126 06:41:36.661204 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 26 06:41:36 crc kubenswrapper[4775]: E1126 06:41:36.661393 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n96hbbhfh56bh555hb4h67dh5fch74h585h54dh676h5b7h5d5h6dh595h696h58fh65bh5f5h57bhb9h577hdh79h569h5f9h544h555h588h5d4h678q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2prm6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7645c8b599-r5tdh_openstack(23563003-3975-4d09-aec4-3b3004aa3672): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:41:36 crc kubenswrapper[4775]: E1126 06:41:36.667090 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7645c8b599-r5tdh" podUID="23563003-3975-4d09-aec4-3b3004aa3672" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.677072 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-config-data\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-combined-ca-bundle\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767629 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-logs\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767645 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm9cd\" (UniqueName: \"kubernetes.io/projected/1d277cfa-56f7-4692-8724-980d3f6dbe0c-kube-api-access-lm9cd\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767704 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-httpd-run\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767738 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-public-tls-certs\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767771 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-scripts\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.767808 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\" (UID: \"1d277cfa-56f7-4692-8724-980d3f6dbe0c\") " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.768677 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.768833 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-logs" (OuterVolumeSpecName: "logs") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.774477 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.780785 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-scripts" (OuterVolumeSpecName: "scripts") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.787224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d277cfa-56f7-4692-8724-980d3f6dbe0c-kube-api-access-lm9cd" (OuterVolumeSpecName: "kube-api-access-lm9cd") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "kube-api-access-lm9cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.819350 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.849657 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-config-data" (OuterVolumeSpecName: "config-data") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.858144 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1d277cfa-56f7-4692-8724-980d3f6dbe0c" (UID: "1d277cfa-56f7-4692-8724-980d3f6dbe0c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870541 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870604 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870619 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870631 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870642 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm9cd\" (UniqueName: \"kubernetes.io/projected/1d277cfa-56f7-4692-8724-980d3f6dbe0c-kube-api-access-lm9cd\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870654 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d277cfa-56f7-4692-8724-980d3f6dbe0c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870664 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.870673 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d277cfa-56f7-4692-8724-980d3f6dbe0c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.900747 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 06:41:36 crc kubenswrapper[4775]: I1126 06:41:36.973324 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.007333 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d277cfa-56f7-4692-8724-980d3f6dbe0c","Type":"ContainerDied","Data":"960ed5f5772ecf1902800838043f952847cd5cef84e9bed7362fa81d18829f1b"} Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.007395 4775 scope.go:117] "RemoveContainer" containerID="5c194b3dc440f7ca36cac1595ee1d8784f44529eb31faf058772202a181c0519" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.007542 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: E1126 06:41:37.010566 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-f8988" podUID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.077541 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.085657 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.092828 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:37 crc kubenswrapper[4775]: E1126 06:41:37.093274 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-httpd" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.093288 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-httpd" Nov 26 06:41:37 crc kubenswrapper[4775]: E1126 06:41:37.093311 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-log" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.093316 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-log" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.093514 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-log" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.093535 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" containerName="glance-httpd" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.097048 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.100195 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.100377 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.129261 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-config-data\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176428 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-logs\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176475 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176530 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176638 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27g4s\" (UniqueName: \"kubernetes.io/projected/3883f87e-8234-4813-a3b9-774f26e1b2ea-kube-api-access-27g4s\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-scripts\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176722 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.176804 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27g4s\" (UniqueName: \"kubernetes.io/projected/3883f87e-8234-4813-a3b9-774f26e1b2ea-kube-api-access-27g4s\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-scripts\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278900 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278931 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278963 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-config-data\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.278989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-logs\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.279008 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.279220 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.279733 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.279852 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-logs\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.283126 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.283497 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.285275 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-scripts\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.293074 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-config-data\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.295329 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27g4s\" (UniqueName: \"kubernetes.io/projected/3883f87e-8234-4813-a3b9-774f26e1b2ea-kube-api-access-27g4s\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: E1126 06:41:37.295592 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 26 06:41:37 crc kubenswrapper[4775]: E1126 06:41:37.295808 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n86h68bh68bh54ch694h575hb9hc9h66bh78h6h5b5h5cbh5f9h688h59hd8h5b6hfch675h84h95h74h67ch694h645hdh58fh669h5d8h567h669q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5wwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(a14e76fa-5315-4564-9430-9231a8e31a01): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.306983 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " pod="openstack/glance-default-external-api-0" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.337696 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d277cfa-56f7-4692-8724-980d3f6dbe0c" path="/var/lib/kubelet/pods/1d277cfa-56f7-4692-8724-980d3f6dbe0c/volumes" Nov 26 06:41:37 crc kubenswrapper[4775]: I1126 06:41:37.429381 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:41:39 crc kubenswrapper[4775]: E1126 06:41:39.221011 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Nov 26 06:41:39 crc kubenswrapper[4775]: E1126 06:41:39.222247 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5h54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-n9tbr_openstack(b0f163bf-e900-4d6e-b5be-948f927c7b86): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:41:39 crc kubenswrapper[4775]: E1126 06:41:39.223828 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-n9tbr" podUID="b0f163bf-e900-4d6e-b5be-948f927c7b86" Nov 26 06:41:39 crc kubenswrapper[4775]: E1126 06:41:39.241347 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Nov 26 06:41:39 crc kubenswrapper[4775]: E1126 06:41:39.241567 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nb5hbchcfh579h5bdh5ddh546h598h554h555h698h69h579hcbh68h7h56fh558h84h6ch696h576h5f9h57ch69h567h645h545h5ch64fhd9h5f4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tzfbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-f8db555df-xs5d8_openstack(b0a4cb55-482a-4be2-9c7b-48874b6cbbac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:41:39 crc kubenswrapper[4775]: E1126 06:41:39.245654 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-f8db555df-xs5d8" podUID="b0a4cb55-482a-4be2-9c7b-48874b6cbbac" Nov 26 06:41:40 crc kubenswrapper[4775]: E1126 06:41:40.059053 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-n9tbr" podUID="b0f163bf-e900-4d6e-b5be-948f927c7b86" Nov 26 06:41:41 crc kubenswrapper[4775]: I1126 06:41:41.331700 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Nov 26 06:41:41 crc kubenswrapper[4775]: I1126 06:41:41.343749 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.415906 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.471030 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.478499 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.493659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-fernet-keys\") pod \"d97295cb-cfff-4626-b1d0-0935fb711fe0\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.493809 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-combined-ca-bundle\") pod \"d97295cb-cfff-4626-b1d0-0935fb711fe0\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.493888 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-scripts\") pod \"d97295cb-cfff-4626-b1d0-0935fb711fe0\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.493959 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvfd8\" (UniqueName: \"kubernetes.io/projected/d97295cb-cfff-4626-b1d0-0935fb711fe0-kube-api-access-nvfd8\") pod \"d97295cb-cfff-4626-b1d0-0935fb711fe0\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.494030 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-config-data\") pod \"d97295cb-cfff-4626-b1d0-0935fb711fe0\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.494053 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-credential-keys\") pod \"d97295cb-cfff-4626-b1d0-0935fb711fe0\" (UID: \"d97295cb-cfff-4626-b1d0-0935fb711fe0\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.495295 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.500066 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d97295cb-cfff-4626-b1d0-0935fb711fe0" (UID: "d97295cb-cfff-4626-b1d0-0935fb711fe0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.500957 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-scripts" (OuterVolumeSpecName: "scripts") pod "d97295cb-cfff-4626-b1d0-0935fb711fe0" (UID: "d97295cb-cfff-4626-b1d0-0935fb711fe0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.503011 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d97295cb-cfff-4626-b1d0-0935fb711fe0" (UID: "d97295cb-cfff-4626-b1d0-0935fb711fe0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.547006 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97295cb-cfff-4626-b1d0-0935fb711fe0-kube-api-access-nvfd8" (OuterVolumeSpecName: "kube-api-access-nvfd8") pod "d97295cb-cfff-4626-b1d0-0935fb711fe0" (UID: "d97295cb-cfff-4626-b1d0-0935fb711fe0"). InnerVolumeSpecName "kube-api-access-nvfd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.570424 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-config-data" (OuterVolumeSpecName: "config-data") pod "d97295cb-cfff-4626-b1d0-0935fb711fe0" (UID: "d97295cb-cfff-4626-b1d0-0935fb711fe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.575544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d97295cb-cfff-4626-b1d0-0935fb711fe0" (UID: "d97295cb-cfff-4626-b1d0-0935fb711fe0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595285 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-httpd-run\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595326 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-scripts\") pod \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-config-data\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595441 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23563003-3975-4d09-aec4-3b3004aa3672-horizon-secret-key\") pod \"23563003-3975-4d09-aec4-3b3004aa3672\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595459 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-scripts\") pod \"23563003-3975-4d09-aec4-3b3004aa3672\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595480 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79m8j\" (UniqueName: \"kubernetes.io/projected/31d04adf-82d4-4643-97df-8f6d9672b4cb-kube-api-access-79m8j\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-logs\") pod \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595591 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-combined-ca-bundle\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595607 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-internal-tls-certs\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23563003-3975-4d09-aec4-3b3004aa3672-logs\") pod \"23563003-3975-4d09-aec4-3b3004aa3672\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595649 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-config-data\") pod \"23563003-3975-4d09-aec4-3b3004aa3672\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595668 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-logs\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595687 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzfbs\" (UniqueName: \"kubernetes.io/projected/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-kube-api-access-tzfbs\") pod \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595703 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-scripts\") pod \"31d04adf-82d4-4643-97df-8f6d9672b4cb\" (UID: \"31d04adf-82d4-4643-97df-8f6d9672b4cb\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595743 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-config-data\") pod \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595766 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2prm6\" (UniqueName: \"kubernetes.io/projected/23563003-3975-4d09-aec4-3b3004aa3672-kube-api-access-2prm6\") pod \"23563003-3975-4d09-aec4-3b3004aa3672\" (UID: \"23563003-3975-4d09-aec4-3b3004aa3672\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595785 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-horizon-secret-key\") pod \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\" (UID: \"b0a4cb55-482a-4be2-9c7b-48874b6cbbac\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.595863 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596092 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596102 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-logs" (OuterVolumeSpecName: "logs") pod "b0a4cb55-482a-4be2-9c7b-48874b6cbbac" (UID: "b0a4cb55-482a-4be2-9c7b-48874b6cbbac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596111 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596147 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596164 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvfd8\" (UniqueName: \"kubernetes.io/projected/d97295cb-cfff-4626-b1d0-0935fb711fe0-kube-api-access-nvfd8\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596177 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596190 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d97295cb-cfff-4626-b1d0-0935fb711fe0-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596261 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-scripts" (OuterVolumeSpecName: "scripts") pod "b0a4cb55-482a-4be2-9c7b-48874b6cbbac" (UID: "b0a4cb55-482a-4be2-9c7b-48874b6cbbac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.596867 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-scripts" (OuterVolumeSpecName: "scripts") pod "23563003-3975-4d09-aec4-3b3004aa3672" (UID: "23563003-3975-4d09-aec4-3b3004aa3672"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.597097 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-config-data" (OuterVolumeSpecName: "config-data") pod "23563003-3975-4d09-aec4-3b3004aa3672" (UID: "23563003-3975-4d09-aec4-3b3004aa3672"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.597607 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23563003-3975-4d09-aec4-3b3004aa3672-logs" (OuterVolumeSpecName: "logs") pod "23563003-3975-4d09-aec4-3b3004aa3672" (UID: "23563003-3975-4d09-aec4-3b3004aa3672"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.598486 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-logs" (OuterVolumeSpecName: "logs") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.598535 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23563003-3975-4d09-aec4-3b3004aa3672-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "23563003-3975-4d09-aec4-3b3004aa3672" (UID: "23563003-3975-4d09-aec4-3b3004aa3672"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.598676 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-config-data" (OuterVolumeSpecName: "config-data") pod "b0a4cb55-482a-4be2-9c7b-48874b6cbbac" (UID: "b0a4cb55-482a-4be2-9c7b-48874b6cbbac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.611876 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.611890 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b0a4cb55-482a-4be2-9c7b-48874b6cbbac" (UID: "b0a4cb55-482a-4be2-9c7b-48874b6cbbac"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.611882 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d04adf-82d4-4643-97df-8f6d9672b4cb-kube-api-access-79m8j" (OuterVolumeSpecName: "kube-api-access-79m8j") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "kube-api-access-79m8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.611920 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-kube-api-access-tzfbs" (OuterVolumeSpecName: "kube-api-access-tzfbs") pod "b0a4cb55-482a-4be2-9c7b-48874b6cbbac" (UID: "b0a4cb55-482a-4be2-9c7b-48874b6cbbac"). InnerVolumeSpecName "kube-api-access-tzfbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.612079 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-scripts" (OuterVolumeSpecName: "scripts") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.615026 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23563003-3975-4d09-aec4-3b3004aa3672-kube-api-access-2prm6" (OuterVolumeSpecName: "kube-api-access-2prm6") pod "23563003-3975-4d09-aec4-3b3004aa3672" (UID: "23563003-3975-4d09-aec4-3b3004aa3672"). InnerVolumeSpecName "kube-api-access-2prm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.631639 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.645971 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.649912 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-config-data" (OuterVolumeSpecName: "config-data") pod "31d04adf-82d4-4643-97df-8f6d9672b4cb" (UID: "31d04adf-82d4-4643-97df-8f6d9672b4cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697853 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzfbs\" (UniqueName: \"kubernetes.io/projected/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-kube-api-access-tzfbs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697880 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697889 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697899 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2prm6\" (UniqueName: \"kubernetes.io/projected/23563003-3975-4d09-aec4-3b3004aa3672-kube-api-access-2prm6\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697907 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697917 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697925 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697933 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/23563003-3975-4d09-aec4-3b3004aa3672-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697941 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697948 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79m8j\" (UniqueName: \"kubernetes.io/projected/31d04adf-82d4-4643-97df-8f6d9672b4cb-kube-api-access-79m8j\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697983 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.697997 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0a4cb55-482a-4be2-9c7b-48874b6cbbac-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.698008 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.698020 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/31d04adf-82d4-4643-97df-8f6d9672b4cb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.698031 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23563003-3975-4d09-aec4-3b3004aa3672-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.698041 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/23563003-3975-4d09-aec4-3b3004aa3672-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.698051 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31d04adf-82d4-4643-97df-8f6d9672b4cb-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.728515 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:46.799008 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.126270 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7645c8b599-r5tdh" event={"ID":"23563003-3975-4d09-aec4-3b3004aa3672","Type":"ContainerDied","Data":"1be8fbe7e407e66dbf62cebfeafe4a4eba89fc648a8e18f6e7b24752a763f6e8"} Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.126358 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7645c8b599-r5tdh" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.131334 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tjjff" event={"ID":"d97295cb-cfff-4626-b1d0-0935fb711fe0","Type":"ContainerDied","Data":"9e972185132634907c02a609cc60744aded18ff02b72ea698730d58b83c5cd25"} Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.131413 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e972185132634907c02a609cc60744aded18ff02b72ea698730d58b83c5cd25" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.131525 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tjjff" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.138944 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8db555df-xs5d8" event={"ID":"b0a4cb55-482a-4be2-9c7b-48874b6cbbac","Type":"ContainerDied","Data":"39ef0677f1dda6eeeb348bc0da9095c1a903d8555c21722e810d2296a7169b50"} Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.139038 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8db555df-xs5d8" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.144119 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"31d04adf-82d4-4643-97df-8f6d9672b4cb","Type":"ContainerDied","Data":"f718906d2408ae959e644438577df8098db1dffbbefdf74f1b8b10c25cb0922e"} Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.144173 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.207088 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7645c8b599-r5tdh"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.271455 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7645c8b599-r5tdh"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.292782 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f8db555df-xs5d8"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.301969 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f8db555df-xs5d8"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.308226 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.316676 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325342 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:47 crc kubenswrapper[4775]: E1126 06:41:47.325658 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97295cb-cfff-4626-b1d0-0935fb711fe0" containerName="keystone-bootstrap" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325672 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97295cb-cfff-4626-b1d0-0935fb711fe0" containerName="keystone-bootstrap" Nov 26 06:41:47 crc kubenswrapper[4775]: E1126 06:41:47.325706 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-httpd" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325730 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-httpd" Nov 26 06:41:47 crc kubenswrapper[4775]: E1126 06:41:47.325746 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-log" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325752 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-log" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325896 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d97295cb-cfff-4626-b1d0-0935fb711fe0" containerName="keystone-bootstrap" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325910 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-log" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.325919 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" containerName="glance-httpd" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.327388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.331760 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.332251 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.340024 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23563003-3975-4d09-aec4-3b3004aa3672" path="/var/lib/kubelet/pods/23563003-3975-4d09-aec4-3b3004aa3672/volumes" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.340781 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d04adf-82d4-4643-97df-8f6d9672b4cb" path="/var/lib/kubelet/pods/31d04adf-82d4-4643-97df-8f6d9672b4cb/volumes" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.341373 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0a4cb55-482a-4be2-9c7b-48874b6cbbac" path="/var/lib/kubelet/pods/b0a4cb55-482a-4be2-9c7b-48874b6cbbac/volumes" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.341689 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.412788 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413044 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7w6d\" (UniqueName: \"kubernetes.io/projected/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-kube-api-access-v7w6d\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413071 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413782 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413838 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.413922 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.487047 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tjjff"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.495412 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tjjff"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.515830 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.515903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7w6d\" (UniqueName: \"kubernetes.io/projected/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-kube-api-access-v7w6d\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.515940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.515982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516004 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516036 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516068 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516120 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516159 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516532 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-logs\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.516592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.521544 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.521568 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.521580 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.529083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.531837 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7w6d\" (UniqueName: \"kubernetes.io/projected/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-kube-api-access-v7w6d\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.545482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.600797 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hsfmj"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.601993 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.607081 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.607293 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.607298 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.607538 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.607612 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jvmv4" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.623841 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hsfmj"] Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.657874 4775 scope.go:117] "RemoveContainer" containerID="d5c7cdda2820e517050f9407209991a6aed7a48b7397a085a2eb82d0a47a5223" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.658897 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:47 crc kubenswrapper[4775]: E1126 06:41:47.679118 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 26 06:41:47 crc kubenswrapper[4775]: E1126 06:41:47.679278 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvt6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-ng8zs_openstack(9d6b7cfb-7b59-4f0b-b36c-94dae19e512d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 06:41:47 crc kubenswrapper[4775]: E1126 06:41:47.681155 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-ng8zs" podUID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.718677 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-scripts\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.718769 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-config-data\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.718800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-fernet-keys\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.718827 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-combined-ca-bundle\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.718851 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlf5v\" (UniqueName: \"kubernetes.io/projected/59435a4c-0feb-4d88-8929-e7c0ea22302d-kube-api-access-vlf5v\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.719106 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-credential-keys\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.728606 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.820211 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-config\") pod \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.820270 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-sb\") pod \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.820456 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx4nr\" (UniqueName: \"kubernetes.io/projected/ed38647f-f4ed-4993-83f1-f7f34255e9b7-kube-api-access-lx4nr\") pod \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.820568 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-dns-svc\") pod \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.820604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-nb\") pod \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\" (UID: \"ed38647f-f4ed-4993-83f1-f7f34255e9b7\") " Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.820910 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-credential-keys\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.821032 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-scripts\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.821082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-config-data\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.821104 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-fernet-keys\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.821131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-combined-ca-bundle\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.821151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlf5v\" (UniqueName: \"kubernetes.io/projected/59435a4c-0feb-4d88-8929-e7c0ea22302d-kube-api-access-vlf5v\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.824074 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-scripts\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.824541 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-fernet-keys\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.825632 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-credential-keys\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.826118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed38647f-f4ed-4993-83f1-f7f34255e9b7-kube-api-access-lx4nr" (OuterVolumeSpecName: "kube-api-access-lx4nr") pod "ed38647f-f4ed-4993-83f1-f7f34255e9b7" (UID: "ed38647f-f4ed-4993-83f1-f7f34255e9b7"). InnerVolumeSpecName "kube-api-access-lx4nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.828355 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-config-data\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.834381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-combined-ca-bundle\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.837612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlf5v\" (UniqueName: \"kubernetes.io/projected/59435a4c-0feb-4d88-8929-e7c0ea22302d-kube-api-access-vlf5v\") pod \"keystone-bootstrap-hsfmj\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.871154 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed38647f-f4ed-4993-83f1-f7f34255e9b7" (UID: "ed38647f-f4ed-4993-83f1-f7f34255e9b7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.883837 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed38647f-f4ed-4993-83f1-f7f34255e9b7" (UID: "ed38647f-f4ed-4993-83f1-f7f34255e9b7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.887224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-config" (OuterVolumeSpecName: "config") pod "ed38647f-f4ed-4993-83f1-f7f34255e9b7" (UID: "ed38647f-f4ed-4993-83f1-f7f34255e9b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.887610 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed38647f-f4ed-4993-83f1-f7f34255e9b7" (UID: "ed38647f-f4ed-4993-83f1-f7f34255e9b7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.923217 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.923296 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx4nr\" (UniqueName: \"kubernetes.io/projected/ed38647f-f4ed-4993-83f1-f7f34255e9b7-kube-api-access-lx4nr\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.923338 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.923351 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.923362 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:47 crc kubenswrapper[4775]: I1126 06:41:47.923372 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed38647f-f4ed-4993-83f1-f7f34255e9b7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.001117 4775 scope.go:117] "RemoveContainer" containerID="8282484c093c2fc8a5678ed3717256fdb144f4b27a85f2b08f8ffb603b8df6de" Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.056142 4775 scope.go:117] "RemoveContainer" containerID="bb28b80f049fe85c5df266f2d58eb9c20ac620fb7a9b283daed1076a7459f2fc" Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.161455 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.161751 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" event={"ID":"ed38647f-f4ed-4993-83f1-f7f34255e9b7","Type":"ContainerDied","Data":"ecd9b4f002cecb047982e56333b71e2897238a0db7e9427485d008ff060ec81d"} Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.161820 4775 scope.go:117] "RemoveContainer" containerID="e8aebe0fab1e656480dc0057b764996044e1fa9b766e37f0b917ec6ebbed7bcc" Nov 26 06:41:48 crc kubenswrapper[4775]: E1126 06:41:48.166686 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-ng8zs" podUID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.210499 4775 scope.go:117] "RemoveContainer" containerID="42062ad7ec7b4257800732aef18d6bd9d02d0261bc387d67ac6eecda9946f547" Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.233526 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jggzv"] Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.239224 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-jggzv"] Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.369445 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-9c9599df6-zd8gh"] Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.505710 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5886fff4d-zcj7s"] Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.599252 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.608564 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hsfmj"] Nov 26 06:41:48 crc kubenswrapper[4775]: I1126 06:41:48.699465 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:41:48 crc kubenswrapper[4775]: W1126 06:41:48.705774 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf1a7acd_833b_439f_bcb9_f83dae6dcf53.slice/crio-24c95b01c91617d9af139ddc8b1200b259e1152788a6198910d2d3a225858b4a WatchSource:0}: Error finding container 24c95b01c91617d9af139ddc8b1200b259e1152788a6198910d2d3a225858b4a: Status 404 returned error can't find the container with id 24c95b01c91617d9af139ddc8b1200b259e1152788a6198910d2d3a225858b4a Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.176498 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3883f87e-8234-4813-a3b9-774f26e1b2ea","Type":"ContainerStarted","Data":"27332bf08d0f217b4aa0312e4ce04d0256c4721a3e03bbe63d605d4a7713144b"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.178705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerStarted","Data":"ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.180934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c9599df6-zd8gh" event={"ID":"f578a3d5-f846-44b6-bc2b-3d98188dffff","Type":"ContainerStarted","Data":"daeb29f5b84018afd04e36369924d6cd4e02a84131d068277215152ba8a97f09"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.180957 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c9599df6-zd8gh" event={"ID":"f578a3d5-f846-44b6-bc2b-3d98188dffff","Type":"ContainerStarted","Data":"4d9fa6b29a571cca3a7fdc8e853ea59e73b0de6fb2383268ab0d1da0b2fdd390"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.180966 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-9c9599df6-zd8gh" event={"ID":"f578a3d5-f846-44b6-bc2b-3d98188dffff","Type":"ContainerStarted","Data":"95af14f9c63cb75ba2b9f3a0cfe195a3245744f880a8bb768b0e7fbc634c5633"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.183244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsfmj" event={"ID":"59435a4c-0feb-4d88-8929-e7c0ea22302d","Type":"ContainerStarted","Data":"b774e3ccc4b186bdb6017dfe69399898f2e6696671dc92422bacfa37e4d3fd85"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.183267 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsfmj" event={"ID":"59435a4c-0feb-4d88-8929-e7c0ea22302d","Type":"ContainerStarted","Data":"e1008a074b3a697d8f3bb8d913cf04ebaa8b2b8276cc139a6bbf5a665947372a"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.187401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5886fff4d-zcj7s" event={"ID":"fb607340-2aa9-4b57-a4b7-a6a64457cae3","Type":"ContainerStarted","Data":"4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.187427 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5886fff4d-zcj7s" event={"ID":"fb607340-2aa9-4b57-a4b7-a6a64457cae3","Type":"ContainerStarted","Data":"aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.187437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5886fff4d-zcj7s" event={"ID":"fb607340-2aa9-4b57-a4b7-a6a64457cae3","Type":"ContainerStarted","Data":"13cab573fa3085fbc1a2ee95bad6d2f475a0e4cdf60ff95d53d3e4585bc1541c"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.190483 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84989d94bf-7vmkz" event={"ID":"10c96169-294e-4273-8e08-92c677dfc01c","Type":"ContainerStarted","Data":"0d33e092249e992f336e7facec4df4d9ffea73cb6fe83a88efd5ff023dbb300f"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.190525 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84989d94bf-7vmkz" event={"ID":"10c96169-294e-4273-8e08-92c677dfc01c","Type":"ContainerStarted","Data":"cbb5041fedb73961d9aac8c4248c3cdd509372660fff75d643233a535805de73"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.190627 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84989d94bf-7vmkz" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon-log" containerID="cri-o://cbb5041fedb73961d9aac8c4248c3cdd509372660fff75d643233a535805de73" gracePeriod=30 Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.190894 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-84989d94bf-7vmkz" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon" containerID="cri-o://0d33e092249e992f336e7facec4df4d9ffea73cb6fe83a88efd5ff023dbb300f" gracePeriod=30 Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.193942 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf1a7acd-833b-439f-bcb9-f83dae6dcf53","Type":"ContainerStarted","Data":"24c95b01c91617d9af139ddc8b1200b259e1152788a6198910d2d3a225858b4a"} Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.216516 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-9c9599df6-zd8gh" podStartSLOduration=21.216496711 podStartE2EDuration="21.216496711s" podCreationTimestamp="2025-11-26 06:41:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:49.198280707 +0000 UTC m=+992.559584659" watchObservedRunningTime="2025-11-26 06:41:49.216496711 +0000 UTC m=+992.577800673" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.223551 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-84989d94bf-7vmkz" podStartSLOduration=4.37101742 podStartE2EDuration="27.223532444s" podCreationTimestamp="2025-11-26 06:41:22 +0000 UTC" firstStartedPulling="2025-11-26 06:41:23.459269407 +0000 UTC m=+966.820573349" lastFinishedPulling="2025-11-26 06:41:46.311784421 +0000 UTC m=+989.673088373" observedRunningTime="2025-11-26 06:41:49.218314778 +0000 UTC m=+992.579618730" watchObservedRunningTime="2025-11-26 06:41:49.223532444 +0000 UTC m=+992.584836396" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.254677 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hsfmj" podStartSLOduration=2.254656663 podStartE2EDuration="2.254656663s" podCreationTimestamp="2025-11-26 06:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:49.24761908 +0000 UTC m=+992.608923032" watchObservedRunningTime="2025-11-26 06:41:49.254656663 +0000 UTC m=+992.615960615" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.276155 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5886fff4d-zcj7s" podStartSLOduration=21.276133081 podStartE2EDuration="21.276133081s" podCreationTimestamp="2025-11-26 06:41:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:49.266384328 +0000 UTC m=+992.627688290" watchObservedRunningTime="2025-11-26 06:41:49.276133081 +0000 UTC m=+992.637437043" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.288584 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.288660 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.347320 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d97295cb-cfff-4626-b1d0-0935fb711fe0" path="/var/lib/kubelet/pods/d97295cb-cfff-4626-b1d0-0935fb711fe0/volumes" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.349528 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" path="/var/lib/kubelet/pods/ed38647f-f4ed-4993-83f1-f7f34255e9b7/volumes" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.379824 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:49 crc kubenswrapper[4775]: I1126 06:41:49.379872 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:41:50 crc kubenswrapper[4775]: I1126 06:41:50.219419 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf1a7acd-833b-439f-bcb9-f83dae6dcf53","Type":"ContainerStarted","Data":"c4063b48c0ec06f1e596cb6746b388b87519e7665b6fe1c57927aead4c5db05d"} Nov 26 06:41:50 crc kubenswrapper[4775]: I1126 06:41:50.222977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3883f87e-8234-4813-a3b9-774f26e1b2ea","Type":"ContainerStarted","Data":"a485eab588369566aa8f255e9ec6ef51f4510c4c10250579ae1223f55acd0df8"} Nov 26 06:41:51 crc kubenswrapper[4775]: I1126 06:41:51.331074 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-jggzv" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Nov 26 06:41:52 crc kubenswrapper[4775]: I1126 06:41:52.722602 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:41:53 crc kubenswrapper[4775]: I1126 06:41:53.275859 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf1a7acd-833b-439f-bcb9-f83dae6dcf53","Type":"ContainerStarted","Data":"d93bd4dedbb36b7574fdca052093e0f1cb8b51986765480b9a021da0ea08de96"} Nov 26 06:41:53 crc kubenswrapper[4775]: I1126 06:41:53.282192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3883f87e-8234-4813-a3b9-774f26e1b2ea","Type":"ContainerStarted","Data":"778db2e53ba426fd87d3a85924c15d4815c08f416f191dd9dca0fc0dfd65600a"} Nov 26 06:41:53 crc kubenswrapper[4775]: I1126 06:41:53.312003 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=16.311969347 podStartE2EDuration="16.311969347s" podCreationTimestamp="2025-11-26 06:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:53.30787354 +0000 UTC m=+996.669177502" watchObservedRunningTime="2025-11-26 06:41:53.311969347 +0000 UTC m=+996.673273359" Nov 26 06:41:54 crc kubenswrapper[4775]: I1126 06:41:54.319901 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.319879809 podStartE2EDuration="7.319879809s" podCreationTimestamp="2025-11-26 06:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:41:54.317916448 +0000 UTC m=+997.679220400" watchObservedRunningTime="2025-11-26 06:41:54.319879809 +0000 UTC m=+997.681183781" Nov 26 06:41:55 crc kubenswrapper[4775]: I1126 06:41:55.311988 4775 generic.go:334] "Generic (PLEG): container finished" podID="747654d0-4f51-411e-958b-34f3cd5403eb" containerID="b13fda92c3c8312d7063282c76057773adc40ebd3ea6936dc6aebb2f25e040ec" exitCode=0 Nov 26 06:41:55 crc kubenswrapper[4775]: I1126 06:41:55.312138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqh7n" event={"ID":"747654d0-4f51-411e-958b-34f3cd5403eb","Type":"ContainerDied","Data":"b13fda92c3c8312d7063282c76057773adc40ebd3ea6936dc6aebb2f25e040ec"} Nov 26 06:41:55 crc kubenswrapper[4775]: I1126 06:41:55.314756 4775 generic.go:334] "Generic (PLEG): container finished" podID="59435a4c-0feb-4d88-8929-e7c0ea22302d" containerID="b774e3ccc4b186bdb6017dfe69399898f2e6696671dc92422bacfa37e4d3fd85" exitCode=0 Nov 26 06:41:55 crc kubenswrapper[4775]: I1126 06:41:55.314812 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsfmj" event={"ID":"59435a4c-0feb-4d88-8929-e7c0ea22302d","Type":"ContainerDied","Data":"b774e3ccc4b186bdb6017dfe69399898f2e6696671dc92422bacfa37e4d3fd85"} Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.430195 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.430548 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.483088 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.483696 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.660127 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.660433 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.696194 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.709844 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.789812 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.822609 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.834341 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-combined-ca-bundle\") pod \"747654d0-4f51-411e-958b-34f3cd5403eb\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.835308 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-fernet-keys\") pod \"59435a4c-0feb-4d88-8929-e7c0ea22302d\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.835442 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlf5v\" (UniqueName: \"kubernetes.io/projected/59435a4c-0feb-4d88-8929-e7c0ea22302d-kube-api-access-vlf5v\") pod \"59435a4c-0feb-4d88-8929-e7c0ea22302d\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.836359 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-config-data\") pod \"59435a4c-0feb-4d88-8929-e7c0ea22302d\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.836396 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-config\") pod \"747654d0-4f51-411e-958b-34f3cd5403eb\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.836430 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-combined-ca-bundle\") pod \"59435a4c-0feb-4d88-8929-e7c0ea22302d\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.836495 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nntvk\" (UniqueName: \"kubernetes.io/projected/747654d0-4f51-411e-958b-34f3cd5403eb-kube-api-access-nntvk\") pod \"747654d0-4f51-411e-958b-34f3cd5403eb\" (UID: \"747654d0-4f51-411e-958b-34f3cd5403eb\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.836565 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-scripts\") pod \"59435a4c-0feb-4d88-8929-e7c0ea22302d\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.836631 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-credential-keys\") pod \"59435a4c-0feb-4d88-8929-e7c0ea22302d\" (UID: \"59435a4c-0feb-4d88-8929-e7c0ea22302d\") " Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.844277 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59435a4c-0feb-4d88-8929-e7c0ea22302d-kube-api-access-vlf5v" (OuterVolumeSpecName: "kube-api-access-vlf5v") pod "59435a4c-0feb-4d88-8929-e7c0ea22302d" (UID: "59435a4c-0feb-4d88-8929-e7c0ea22302d"). InnerVolumeSpecName "kube-api-access-vlf5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.846126 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-scripts" (OuterVolumeSpecName: "scripts") pod "59435a4c-0feb-4d88-8929-e7c0ea22302d" (UID: "59435a4c-0feb-4d88-8929-e7c0ea22302d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.850142 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "59435a4c-0feb-4d88-8929-e7c0ea22302d" (UID: "59435a4c-0feb-4d88-8929-e7c0ea22302d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.850153 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "59435a4c-0feb-4d88-8929-e7c0ea22302d" (UID: "59435a4c-0feb-4d88-8929-e7c0ea22302d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.859297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/747654d0-4f51-411e-958b-34f3cd5403eb-kube-api-access-nntvk" (OuterVolumeSpecName: "kube-api-access-nntvk") pod "747654d0-4f51-411e-958b-34f3cd5403eb" (UID: "747654d0-4f51-411e-958b-34f3cd5403eb"). InnerVolumeSpecName "kube-api-access-nntvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.938586 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nntvk\" (UniqueName: \"kubernetes.io/projected/747654d0-4f51-411e-958b-34f3cd5403eb-kube-api-access-nntvk\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.938622 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.938635 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.938647 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.938657 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlf5v\" (UniqueName: \"kubernetes.io/projected/59435a4c-0feb-4d88-8929-e7c0ea22302d-kube-api-access-vlf5v\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.944499 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-config-data" (OuterVolumeSpecName: "config-data") pod "59435a4c-0feb-4d88-8929-e7c0ea22302d" (UID: "59435a4c-0feb-4d88-8929-e7c0ea22302d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.947166 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "747654d0-4f51-411e-958b-34f3cd5403eb" (UID: "747654d0-4f51-411e-958b-34f3cd5403eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.949859 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-config" (OuterVolumeSpecName: "config") pod "747654d0-4f51-411e-958b-34f3cd5403eb" (UID: "747654d0-4f51-411e-958b-34f3cd5403eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:57 crc kubenswrapper[4775]: I1126 06:41:57.952524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59435a4c-0feb-4d88-8929-e7c0ea22302d" (UID: "59435a4c-0feb-4d88-8929-e7c0ea22302d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.045970 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.046006 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.046016 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59435a4c-0feb-4d88-8929-e7c0ea22302d-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.046025 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/747654d0-4f51-411e-958b-34f3cd5403eb-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.345326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8988" event={"ID":"25784e47-2e5a-4df0-9aa2-c2a42659cb40","Type":"ContainerStarted","Data":"e253f7dee08463f3f7bac7c6e2789f784f447114cc5764ee8b230022d8fca446"} Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.348869 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerStarted","Data":"057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23"} Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.352385 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-mqh7n" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.352656 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-mqh7n" event={"ID":"747654d0-4f51-411e-958b-34f3cd5403eb","Type":"ContainerDied","Data":"07b4e70c36ff7fcff560dc391cb74e581def5583117ac13a90adbef3f460cac1"} Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.352698 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07b4e70c36ff7fcff560dc391cb74e581def5583117ac13a90adbef3f460cac1" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.365509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n9tbr" event={"ID":"b0f163bf-e900-4d6e-b5be-948f927c7b86","Type":"ContainerStarted","Data":"e5afd9d609d560b067e22b56f0325151bf9412988dfc9ca695c325edba363d09"} Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.377751 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsfmj" event={"ID":"59435a4c-0feb-4d88-8929-e7c0ea22302d","Type":"ContainerDied","Data":"e1008a074b3a697d8f3bb8d913cf04ebaa8b2b8276cc139a6bbf5a665947372a"} Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.377811 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1008a074b3a697d8f3bb8d913cf04ebaa8b2b8276cc139a6bbf5a665947372a" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.378042 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsfmj" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.381154 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.381955 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.381982 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.382293 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.386935 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-f8988" podStartSLOduration=2.469163489 podStartE2EDuration="38.386912115s" podCreationTimestamp="2025-11-26 06:41:20 +0000 UTC" firstStartedPulling="2025-11-26 06:41:21.608543366 +0000 UTC m=+964.969847318" lastFinishedPulling="2025-11-26 06:41:57.526291952 +0000 UTC m=+1000.887595944" observedRunningTime="2025-11-26 06:41:58.365410816 +0000 UTC m=+1001.726714778" watchObservedRunningTime="2025-11-26 06:41:58.386912115 +0000 UTC m=+1001.748216087" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.395649 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-n9tbr" podStartSLOduration=2.862378447 podStartE2EDuration="38.395633182s" podCreationTimestamp="2025-11-26 06:41:20 +0000 UTC" firstStartedPulling="2025-11-26 06:41:22.099144326 +0000 UTC m=+965.460448278" lastFinishedPulling="2025-11-26 06:41:57.632399061 +0000 UTC m=+1000.993703013" observedRunningTime="2025-11-26 06:41:58.383073335 +0000 UTC m=+1001.744377297" watchObservedRunningTime="2025-11-26 06:41:58.395633182 +0000 UTC m=+1001.756937154" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.893808 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7679ff5658-lwrpm"] Nov 26 06:41:58 crc kubenswrapper[4775]: E1126 06:41:58.894613 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="init" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.894635 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="init" Nov 26 06:41:58 crc kubenswrapper[4775]: E1126 06:41:58.894655 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.894665 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" Nov 26 06:41:58 crc kubenswrapper[4775]: E1126 06:41:58.894693 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="747654d0-4f51-411e-958b-34f3cd5403eb" containerName="neutron-db-sync" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.894717 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="747654d0-4f51-411e-958b-34f3cd5403eb" containerName="neutron-db-sync" Nov 26 06:41:58 crc kubenswrapper[4775]: E1126 06:41:58.894766 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59435a4c-0feb-4d88-8929-e7c0ea22302d" containerName="keystone-bootstrap" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.894780 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="59435a4c-0feb-4d88-8929-e7c0ea22302d" containerName="keystone-bootstrap" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.895085 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed38647f-f4ed-4993-83f1-f7f34255e9b7" containerName="dnsmasq-dns" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.895125 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="59435a4c-0feb-4d88-8929-e7c0ea22302d" containerName="keystone-bootstrap" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.895141 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="747654d0-4f51-411e-958b-34f3cd5403eb" containerName="neutron-db-sync" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.896010 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.898675 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.898770 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jvmv4" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.899135 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.899186 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.899360 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.901931 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 26 06:41:58 crc kubenswrapper[4775]: I1126 06:41:58.918826 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7679ff5658-lwrpm"] Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.070875 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-5twjc"] Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.072657 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079128 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-scripts\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079200 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-public-tls-certs\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-internal-tls-certs\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079402 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-credential-keys\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079453 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx696\" (UniqueName: \"kubernetes.io/projected/f06e8ee2-5844-487f-80e7-cdbba5909f74-kube-api-access-vx696\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079515 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-combined-ca-bundle\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079673 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-config-data\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.079774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-fernet-keys\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.108762 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-5twjc"] Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184164 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-credential-keys\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx696\" (UniqueName: \"kubernetes.io/projected/f06e8ee2-5844-487f-80e7-cdbba5909f74-kube-api-access-vx696\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184299 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-combined-ca-bundle\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184334 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184419 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184453 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-config-data\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184582 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6gs9\" (UniqueName: \"kubernetes.io/projected/fec155eb-aac6-4e4e-a8c7-1487a06671e3-kube-api-access-x6gs9\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184671 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-fernet-keys\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.184782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-scripts\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.185202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-public-tls-certs\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.185247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.185300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-config\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.185374 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.185404 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-internal-tls-certs\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.193165 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-public-tls-certs\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.193563 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-internal-tls-certs\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.195555 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-fernet-keys\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.196961 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-credential-keys\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.197588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-combined-ca-bundle\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.202615 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-scripts\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.209495 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx696\" (UniqueName: \"kubernetes.io/projected/f06e8ee2-5844-487f-80e7-cdbba5909f74-kube-api-access-vx696\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.232933 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5775c97bd4-547ct"] Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.234534 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.238046 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zq8z7" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.238209 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.238413 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.238522 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.254976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f06e8ee2-5844-487f-80e7-cdbba5909f74-config-data\") pod \"keystone-7679ff5658-lwrpm\" (UID: \"f06e8ee2-5844-487f-80e7-cdbba5909f74\") " pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.282831 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5775c97bd4-547ct"] Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-ovndb-tls-certs\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291760 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6gs9\" (UniqueName: \"kubernetes.io/projected/fec155eb-aac6-4e4e-a8c7-1487a06671e3-kube-api-access-x6gs9\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291802 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-config\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-combined-ca-bundle\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291920 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-config\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291961 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.291993 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t67nr\" (UniqueName: \"kubernetes.io/projected/116bcb10-21df-40b7-b110-10206c92ea20-kube-api-access-t67nr\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.292026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-httpd-config\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.292084 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.292151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.294551 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.297563 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.313683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-config\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.313926 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.314315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.323440 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6gs9\" (UniqueName: \"kubernetes.io/projected/fec155eb-aac6-4e4e-a8c7-1487a06671e3-kube-api-access-x6gs9\") pod \"dnsmasq-dns-84b966f6c9-5twjc\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.323554 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.394576 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-ovndb-tls-certs\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.394623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-config\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.394648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-combined-ca-bundle\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.394686 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t67nr\" (UniqueName: \"kubernetes.io/projected/116bcb10-21df-40b7-b110-10206c92ea20-kube-api-access-t67nr\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.394709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-httpd-config\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.400784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-httpd-config\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.404252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.404788 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-9c9599df6-zd8gh" podUID="f578a3d5-f846-44b6-bc2b-3d98188dffff" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.414936 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-config\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.422939 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-ovndb-tls-certs\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.423906 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-combined-ca-bundle\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.450389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t67nr\" (UniqueName: \"kubernetes.io/projected/116bcb10-21df-40b7-b110-10206c92ea20-kube-api-access-t67nr\") pod \"neutron-5775c97bd4-547ct\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.512122 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.596222 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:41:59 crc kubenswrapper[4775]: I1126 06:41:59.992335 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-5twjc"] Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.139696 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7679ff5658-lwrpm"] Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.423433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" event={"ID":"fec155eb-aac6-4e4e-a8c7-1487a06671e3","Type":"ContainerStarted","Data":"8ee82aecb50376f57c5577206da9331339cde15f1db337daa747b6fd8181183d"} Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.428268 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.428265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7679ff5658-lwrpm" event={"ID":"f06e8ee2-5844-487f-80e7-cdbba5909f74","Type":"ContainerStarted","Data":"2e7cfe6fe02f0dbf1e870162b3bc2897e99c6752c4bee1c3d009e4a8d4c83949"} Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.428290 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.883498 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.883803 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.892521 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 26 06:42:00 crc kubenswrapper[4775]: I1126 06:42:00.908442 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5775c97bd4-547ct"] Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.440597 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.464249 4775 generic.go:334] "Generic (PLEG): container finished" podID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerID="498df95d523b87b0f6e1f4d8f7dd33310c768ef1d0e2de356d0814794f78fec6" exitCode=0 Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.464304 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" event={"ID":"fec155eb-aac6-4e4e-a8c7-1487a06671e3","Type":"ContainerDied","Data":"498df95d523b87b0f6e1f4d8f7dd33310c768ef1d0e2de356d0814794f78fec6"} Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.480569 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7679ff5658-lwrpm" event={"ID":"f06e8ee2-5844-487f-80e7-cdbba5909f74","Type":"ContainerStarted","Data":"7fbd0951cbdafc96be71e0c43abf4c9291b6f44e7648421b614625e764886fab"} Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.481334 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.492757 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ng8zs" event={"ID":"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d","Type":"ContainerStarted","Data":"312f44c48e89b1dc67f8ba575f2c8bd12f5c18407f3569f8b5a705f41a8a3211"} Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.511537 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.511729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5775c97bd4-547ct" event={"ID":"116bcb10-21df-40b7-b110-10206c92ea20","Type":"ContainerStarted","Data":"b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91"} Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.511761 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5775c97bd4-547ct" event={"ID":"116bcb10-21df-40b7-b110-10206c92ea20","Type":"ContainerStarted","Data":"5df1d6ab0c51696cf041d59af53b40fe32eb3fb139b6b321ac85e0997a283111"} Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.518045 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7679ff5658-lwrpm" podStartSLOduration=3.518026581 podStartE2EDuration="3.518026581s" podCreationTimestamp="2025-11-26 06:41:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:01.510206758 +0000 UTC m=+1004.871510730" watchObservedRunningTime="2025-11-26 06:42:01.518026581 +0000 UTC m=+1004.879330533" Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.549151 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-ng8zs" podStartSLOduration=3.9866879219999998 podStartE2EDuration="42.54913405s" podCreationTimestamp="2025-11-26 06:41:19 +0000 UTC" firstStartedPulling="2025-11-26 06:41:21.608124535 +0000 UTC m=+964.969428487" lastFinishedPulling="2025-11-26 06:42:00.170570663 +0000 UTC m=+1003.531874615" observedRunningTime="2025-11-26 06:42:01.543156334 +0000 UTC m=+1004.904460286" watchObservedRunningTime="2025-11-26 06:42:01.54913405 +0000 UTC m=+1004.910438002" Nov 26 06:42:01 crc kubenswrapper[4775]: I1126 06:42:01.647480 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.156901 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54d985cc65-zsv4d"] Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.158612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.161183 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.161636 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.168094 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54d985cc65-zsv4d"] Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285665 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-ovndb-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285791 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-httpd-config\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44xwz\" (UniqueName: \"kubernetes.io/projected/e33bc539-83a2-4077-a430-4872e8587023-kube-api-access-44xwz\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285835 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-combined-ca-bundle\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285869 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-internal-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-public-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.285917 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-config\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-ovndb-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-httpd-config\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44xwz\" (UniqueName: \"kubernetes.io/projected/e33bc539-83a2-4077-a430-4872e8587023-kube-api-access-44xwz\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-combined-ca-bundle\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387888 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-internal-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-public-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.387949 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-config\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.393660 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-public-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.394183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-httpd-config\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.394548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-ovndb-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.399288 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-config\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.401650 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-combined-ca-bundle\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.405267 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e33bc539-83a2-4077-a430-4872e8587023-internal-tls-certs\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.412614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44xwz\" (UniqueName: \"kubernetes.io/projected/e33bc539-83a2-4077-a430-4872e8587023-kube-api-access-44xwz\") pod \"neutron-54d985cc65-zsv4d\" (UID: \"e33bc539-83a2-4077-a430-4872e8587023\") " pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.480698 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.531034 4775 generic.go:334] "Generic (PLEG): container finished" podID="b0f163bf-e900-4d6e-b5be-948f927c7b86" containerID="e5afd9d609d560b067e22b56f0325151bf9412988dfc9ca695c325edba363d09" exitCode=0 Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.531107 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n9tbr" event={"ID":"b0f163bf-e900-4d6e-b5be-948f927c7b86","Type":"ContainerDied","Data":"e5afd9d609d560b067e22b56f0325151bf9412988dfc9ca695c325edba363d09"} Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.534991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5775c97bd4-547ct" event={"ID":"116bcb10-21df-40b7-b110-10206c92ea20","Type":"ContainerStarted","Data":"c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0"} Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.535212 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.537934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" event={"ID":"fec155eb-aac6-4e4e-a8c7-1487a06671e3","Type":"ContainerStarted","Data":"0aa3eaaabe805177979bb57b4dd85eedb48e6cd8e766aa9381e386554ac49ff8"} Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.538757 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.540128 4775 generic.go:334] "Generic (PLEG): container finished" podID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" containerID="e253f7dee08463f3f7bac7c6e2789f784f447114cc5764ee8b230022d8fca446" exitCode=0 Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.540591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8988" event={"ID":"25784e47-2e5a-4df0-9aa2-c2a42659cb40","Type":"ContainerDied","Data":"e253f7dee08463f3f7bac7c6e2789f784f447114cc5764ee8b230022d8fca446"} Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.584089 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" podStartSLOduration=3.584075264 podStartE2EDuration="3.584075264s" podCreationTimestamp="2025-11-26 06:41:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:02.581319713 +0000 UTC m=+1005.942623675" watchObservedRunningTime="2025-11-26 06:42:02.584075264 +0000 UTC m=+1005.945379206" Nov 26 06:42:02 crc kubenswrapper[4775]: I1126 06:42:02.613755 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5775c97bd4-547ct" podStartSLOduration=3.613733886 podStartE2EDuration="3.613733886s" podCreationTimestamp="2025-11-26 06:41:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:02.60007127 +0000 UTC m=+1005.961375222" watchObservedRunningTime="2025-11-26 06:42:02.613733886 +0000 UTC m=+1005.975037838" Nov 26 06:42:03 crc kubenswrapper[4775]: I1126 06:42:03.146712 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54d985cc65-zsv4d"] Nov 26 06:42:03 crc kubenswrapper[4775]: W1126 06:42:03.159862 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode33bc539_83a2_4077_a430_4872e8587023.slice/crio-36b63b781cd4fa98945e148f52c9c1c1f8fadda7d1dd1ac40a8b6c6614888b22 WatchSource:0}: Error finding container 36b63b781cd4fa98945e148f52c9c1c1f8fadda7d1dd1ac40a8b6c6614888b22: Status 404 returned error can't find the container with id 36b63b781cd4fa98945e148f52c9c1c1f8fadda7d1dd1ac40a8b6c6614888b22 Nov 26 06:42:03 crc kubenswrapper[4775]: I1126 06:42:03.551771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54d985cc65-zsv4d" event={"ID":"e33bc539-83a2-4077-a430-4872e8587023","Type":"ContainerStarted","Data":"a6aa271cbbda8a71fe7b35c7fc7d69f790455520c1b1bbcf10ab45d5029a37fa"} Nov 26 06:42:03 crc kubenswrapper[4775]: I1126 06:42:03.551818 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54d985cc65-zsv4d" event={"ID":"e33bc539-83a2-4077-a430-4872e8587023","Type":"ContainerStarted","Data":"36b63b781cd4fa98945e148f52c9c1c1f8fadda7d1dd1ac40a8b6c6614888b22"} Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.113427 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n9tbr" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.124926 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8988" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.223471 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7765k\" (UniqueName: \"kubernetes.io/projected/25784e47-2e5a-4df0-9aa2-c2a42659cb40-kube-api-access-7765k\") pod \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.223563 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-combined-ca-bundle\") pod \"b0f163bf-e900-4d6e-b5be-948f927c7b86\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.223611 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-scripts\") pod \"b0f163bf-e900-4d6e-b5be-948f927c7b86\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.223661 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5h54\" (UniqueName: \"kubernetes.io/projected/b0f163bf-e900-4d6e-b5be-948f927c7b86-kube-api-access-b5h54\") pod \"b0f163bf-e900-4d6e-b5be-948f927c7b86\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.225313 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f163bf-e900-4d6e-b5be-948f927c7b86-logs\") pod \"b0f163bf-e900-4d6e-b5be-948f927c7b86\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.225415 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-combined-ca-bundle\") pod \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.225483 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-config-data\") pod \"b0f163bf-e900-4d6e-b5be-948f927c7b86\" (UID: \"b0f163bf-e900-4d6e-b5be-948f927c7b86\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.225511 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-db-sync-config-data\") pod \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\" (UID: \"25784e47-2e5a-4df0-9aa2-c2a42659cb40\") " Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.235040 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0f163bf-e900-4d6e-b5be-948f927c7b86-logs" (OuterVolumeSpecName: "logs") pod "b0f163bf-e900-4d6e-b5be-948f927c7b86" (UID: "b0f163bf-e900-4d6e-b5be-948f927c7b86"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.247348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25784e47-2e5a-4df0-9aa2-c2a42659cb40-kube-api-access-7765k" (OuterVolumeSpecName: "kube-api-access-7765k") pod "25784e47-2e5a-4df0-9aa2-c2a42659cb40" (UID: "25784e47-2e5a-4df0-9aa2-c2a42659cb40"). InnerVolumeSpecName "kube-api-access-7765k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.249075 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f163bf-e900-4d6e-b5be-948f927c7b86-kube-api-access-b5h54" (OuterVolumeSpecName: "kube-api-access-b5h54") pod "b0f163bf-e900-4d6e-b5be-948f927c7b86" (UID: "b0f163bf-e900-4d6e-b5be-948f927c7b86"). InnerVolumeSpecName "kube-api-access-b5h54". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.249156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-scripts" (OuterVolumeSpecName: "scripts") pod "b0f163bf-e900-4d6e-b5be-948f927c7b86" (UID: "b0f163bf-e900-4d6e-b5be-948f927c7b86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.255924 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "25784e47-2e5a-4df0-9aa2-c2a42659cb40" (UID: "25784e47-2e5a-4df0-9aa2-c2a42659cb40"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.267569 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0f163bf-e900-4d6e-b5be-948f927c7b86" (UID: "b0f163bf-e900-4d6e-b5be-948f927c7b86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.274297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-config-data" (OuterVolumeSpecName: "config-data") pod "b0f163bf-e900-4d6e-b5be-948f927c7b86" (UID: "b0f163bf-e900-4d6e-b5be-948f927c7b86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.277298 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25784e47-2e5a-4df0-9aa2-c2a42659cb40" (UID: "25784e47-2e5a-4df0-9aa2-c2a42659cb40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328618 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7765k\" (UniqueName: \"kubernetes.io/projected/25784e47-2e5a-4df0-9aa2-c2a42659cb40-kube-api-access-7765k\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328659 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328674 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328685 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5h54\" (UniqueName: \"kubernetes.io/projected/b0f163bf-e900-4d6e-b5be-948f927c7b86-kube-api-access-b5h54\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328696 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f163bf-e900-4d6e-b5be-948f927c7b86-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328707 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328728 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f163bf-e900-4d6e-b5be-948f927c7b86-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.328739 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/25784e47-2e5a-4df0-9aa2-c2a42659cb40-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.564363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-n9tbr" event={"ID":"b0f163bf-e900-4d6e-b5be-948f927c7b86","Type":"ContainerDied","Data":"d074a1f62bff62a1204c7ed11717deeeb1eb68207a9d489917399041e191b0e3"} Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.564404 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d074a1f62bff62a1204c7ed11717deeeb1eb68207a9d489917399041e191b0e3" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.564471 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-n9tbr" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.568737 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54d985cc65-zsv4d" event={"ID":"e33bc539-83a2-4077-a430-4872e8587023","Type":"ContainerStarted","Data":"0bd05334cf0cbbd9a486f3fe70cb7c87c1a43b6b759061259e3acb6a0c7aee83"} Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.569318 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.574095 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f8988" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.575080 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f8988" event={"ID":"25784e47-2e5a-4df0-9aa2-c2a42659cb40","Type":"ContainerDied","Data":"63fad62ec314c274ede68d1ca00678d6c452d6fec1f4bc728e53dc9b5b0b2548"} Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.575116 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63fad62ec314c274ede68d1ca00678d6c452d6fec1f4bc728e53dc9b5b0b2548" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.604411 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54d985cc65-zsv4d" podStartSLOduration=2.604395994 podStartE2EDuration="2.604395994s" podCreationTimestamp="2025-11-26 06:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:04.600884803 +0000 UTC m=+1007.962188765" watchObservedRunningTime="2025-11-26 06:42:04.604395994 +0000 UTC m=+1007.965699946" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.659527 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5cd45b998d-gjx5d"] Nov 26 06:42:04 crc kubenswrapper[4775]: E1126 06:42:04.660009 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" containerName="barbican-db-sync" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.660031 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" containerName="barbican-db-sync" Nov 26 06:42:04 crc kubenswrapper[4775]: E1126 06:42:04.660077 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f163bf-e900-4d6e-b5be-948f927c7b86" containerName="placement-db-sync" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.660086 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f163bf-e900-4d6e-b5be-948f927c7b86" containerName="placement-db-sync" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.660279 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" containerName="barbican-db-sync" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.660318 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f163bf-e900-4d6e-b5be-948f927c7b86" containerName="placement-db-sync" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.661364 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.665578 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zmxnc" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.665702 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.665868 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.665914 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.665576 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.692406 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cd45b998d-gjx5d"] Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742082 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-scripts\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742165 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-internal-tls-certs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742295 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-public-tls-certs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742318 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-combined-ca-bundle\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742418 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-config-data\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742551 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc391c06-ad32-4c18-9670-481f7614e1a3-logs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.742629 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjk4p\" (UniqueName: \"kubernetes.io/projected/bc391c06-ad32-4c18-9670-481f7614e1a3-kube-api-access-zjk4p\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.836602 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-76d59b886b-lsps8"] Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.838262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844070 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-public-tls-certs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-combined-ca-bundle\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-config-data\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844253 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc391c06-ad32-4c18-9670-481f7614e1a3-logs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844313 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjk4p\" (UniqueName: \"kubernetes.io/projected/bc391c06-ad32-4c18-9670-481f7614e1a3-kube-api-access-zjk4p\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-scripts\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.844374 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-internal-tls-certs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.847329 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.847539 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-d888c" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.847691 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.848245 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc391c06-ad32-4c18-9670-481f7614e1a3-logs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.857377 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-combined-ca-bundle\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.861281 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-internal-tls-certs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.863592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-scripts\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.868671 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-config-data\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.884831 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5bdbc86c75-sgb2x"] Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.885211 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc391c06-ad32-4c18-9670-481f7614e1a3-public-tls-certs\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.886269 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.887374 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjk4p\" (UniqueName: \"kubernetes.io/projected/bc391c06-ad32-4c18-9670-481f7614e1a3-kube-api-access-zjk4p\") pod \"placement-5cd45b998d-gjx5d\" (UID: \"bc391c06-ad32-4c18-9670-481f7614e1a3\") " pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.894028 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.916562 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5bdbc86c75-sgb2x"] Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946603 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-config-data\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-combined-ca-bundle\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-config-data-custom\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-config-data-custom\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946799 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b958489-5435-42a4-bff8-577a26c717c0-logs\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946820 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa465504-4f8f-4c22-8f6b-3bb834f303c2-logs\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946860 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-combined-ca-bundle\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-config-data\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946897 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb6jl\" (UniqueName: \"kubernetes.io/projected/fa465504-4f8f-4c22-8f6b-3bb834f303c2-kube-api-access-wb6jl\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.946919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wp6n\" (UniqueName: \"kubernetes.io/projected/2b958489-5435-42a4-bff8-577a26c717c0-kube-api-access-7wp6n\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:04 crc kubenswrapper[4775]: I1126 06:42:04.957621 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-76d59b886b-lsps8"] Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.003668 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-5twjc"] Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.007140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049616 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-combined-ca-bundle\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049663 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-config-data\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049684 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb6jl\" (UniqueName: \"kubernetes.io/projected/fa465504-4f8f-4c22-8f6b-3bb834f303c2-kube-api-access-wb6jl\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049706 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wp6n\" (UniqueName: \"kubernetes.io/projected/2b958489-5435-42a4-bff8-577a26c717c0-kube-api-access-7wp6n\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049749 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-config-data\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-combined-ca-bundle\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049865 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-config-data-custom\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049896 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-config-data-custom\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b958489-5435-42a4-bff8-577a26c717c0-logs\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.049938 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa465504-4f8f-4c22-8f6b-3bb834f303c2-logs\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.050327 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa465504-4f8f-4c22-8f6b-3bb834f303c2-logs\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.057788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b958489-5435-42a4-bff8-577a26c717c0-logs\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.057870 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-4xnx9"] Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.059246 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.071636 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-config-data-custom\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.072112 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-combined-ca-bundle\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.073183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-config-data\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.077680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b958489-5435-42a4-bff8-577a26c717c0-combined-ca-bundle\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.085292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-config-data-custom\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.086252 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa465504-4f8f-4c22-8f6b-3bb834f303c2-config-data\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.090312 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb6jl\" (UniqueName: \"kubernetes.io/projected/fa465504-4f8f-4c22-8f6b-3bb834f303c2-kube-api-access-wb6jl\") pod \"barbican-keystone-listener-76d59b886b-lsps8\" (UID: \"fa465504-4f8f-4c22-8f6b-3bb834f303c2\") " pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.094892 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wp6n\" (UniqueName: \"kubernetes.io/projected/2b958489-5435-42a4-bff8-577a26c717c0-kube-api-access-7wp6n\") pod \"barbican-worker-5bdbc86c75-sgb2x\" (UID: \"2b958489-5435-42a4-bff8-577a26c717c0\") " pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.107783 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-4xnx9"] Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.119652 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5dfb875fdb-zr8j4"] Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.121260 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.124472 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.125803 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dfb875fdb-zr8j4"] Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.150958 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.150998 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhwrd\" (UniqueName: \"kubernetes.io/projected/373e990b-3bca-4e7b-8543-840f86d0f7b5-kube-api-access-hhwrd\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151037 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/373e990b-3bca-4e7b-8543-840f86d0f7b5-logs\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151092 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151175 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n48wb\" (UniqueName: \"kubernetes.io/projected/ac26de53-1eff-4374-9983-dc657215bf63-kube-api-access-n48wb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151195 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-config\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151211 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-combined-ca-bundle\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151238 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.151269 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data-custom\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.251972 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-config\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252013 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-combined-ca-bundle\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data-custom\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252138 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252155 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhwrd\" (UniqueName: \"kubernetes.io/projected/373e990b-3bca-4e7b-8543-840f86d0f7b5-kube-api-access-hhwrd\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252184 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252204 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/373e990b-3bca-4e7b-8543-840f86d0f7b5-logs\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252226 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.252326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n48wb\" (UniqueName: \"kubernetes.io/projected/ac26de53-1eff-4374-9983-dc657215bf63-kube-api-access-n48wb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.253140 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-config\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.256692 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.267033 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/373e990b-3bca-4e7b-8543-840f86d0f7b5-logs\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.267536 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data-custom\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.267767 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-combined-ca-bundle\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.268441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.269562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.271891 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.272626 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhwrd\" (UniqueName: \"kubernetes.io/projected/373e990b-3bca-4e7b-8543-840f86d0f7b5-kube-api-access-hhwrd\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.273929 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n48wb\" (UniqueName: \"kubernetes.io/projected/ac26de53-1eff-4374-9983-dc657215bf63-kube-api-access-n48wb\") pod \"dnsmasq-dns-75c8ddd69c-4xnx9\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.285116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.288221 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data\") pod \"barbican-api-5dfb875fdb-zr8j4\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.295199 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5bdbc86c75-sgb2x" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.483388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.491480 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:05 crc kubenswrapper[4775]: I1126 06:42:05.580119 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="dnsmasq-dns" containerID="cri-o://0aa3eaaabe805177979bb57b4dd85eedb48e6cd8e766aa9381e386554ac49ff8" gracePeriod=10 Nov 26 06:42:06 crc kubenswrapper[4775]: I1126 06:42:06.593487 4775 generic.go:334] "Generic (PLEG): container finished" podID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerID="0aa3eaaabe805177979bb57b4dd85eedb48e6cd8e766aa9381e386554ac49ff8" exitCode=0 Nov 26 06:42:06 crc kubenswrapper[4775]: I1126 06:42:06.593531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" event={"ID":"fec155eb-aac6-4e4e-a8c7-1487a06671e3","Type":"ContainerDied","Data":"0aa3eaaabe805177979bb57b4dd85eedb48e6cd8e766aa9381e386554ac49ff8"} Nov 26 06:42:07 crc kubenswrapper[4775]: I1126 06:42:07.613332 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" containerID="312f44c48e89b1dc67f8ba575f2c8bd12f5c18407f3569f8b5a705f41a8a3211" exitCode=0 Nov 26 06:42:07 crc kubenswrapper[4775]: I1126 06:42:07.613576 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ng8zs" event={"ID":"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d","Type":"ContainerDied","Data":"312f44c48e89b1dc67f8ba575f2c8bd12f5c18407f3569f8b5a705f41a8a3211"} Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.429909 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5fd8c567d4-z5vr4"] Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.431615 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.436109 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.436449 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.457995 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fd8c567d4-z5vr4"] Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.513974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-public-tls-certs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.514037 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-internal-tls-certs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.514087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-combined-ca-bundle\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.514110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1db6f2a6-c3be-4317-a6ce-802242830778-logs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.514151 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-config-data\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.514267 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-config-data-custom\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.514318 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2ftg\" (UniqueName: \"kubernetes.io/projected/1db6f2a6-c3be-4317-a6ce-802242830778-kube-api-access-v2ftg\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-config-data-custom\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615619 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2ftg\" (UniqueName: \"kubernetes.io/projected/1db6f2a6-c3be-4317-a6ce-802242830778-kube-api-access-v2ftg\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615666 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-public-tls-certs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615704 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-internal-tls-certs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615772 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-combined-ca-bundle\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615797 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1db6f2a6-c3be-4317-a6ce-802242830778-logs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.615836 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-config-data\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.623326 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-config-data\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.623464 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1db6f2a6-c3be-4317-a6ce-802242830778-logs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.623640 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-config-data-custom\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.628038 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-internal-tls-certs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.631191 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-public-tls-certs\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.631885 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db6f2a6-c3be-4317-a6ce-802242830778-combined-ca-bundle\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.658370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2ftg\" (UniqueName: \"kubernetes.io/projected/1db6f2a6-c3be-4317-a6ce-802242830778-kube-api-access-v2ftg\") pod \"barbican-api-5fd8c567d4-z5vr4\" (UID: \"1db6f2a6-c3be-4317-a6ce-802242830778\") " pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:08 crc kubenswrapper[4775]: I1126 06:42:08.768422 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.245636 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.291521 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.329541 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-scripts\") pod \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.329625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-combined-ca-bundle\") pod \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.329703 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvt6r\" (UniqueName: \"kubernetes.io/projected/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-kube-api-access-gvt6r\") pod \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.329850 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-etc-machine-id\") pod \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.329905 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-config-data\") pod \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.329978 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-db-sync-config-data\") pod \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\" (UID: \"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d\") " Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.330785 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" (UID: "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.336065 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-kube-api-access-gvt6r" (OuterVolumeSpecName: "kube-api-access-gvt6r") pod "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" (UID: "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d"). InnerVolumeSpecName "kube-api-access-gvt6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.354320 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" (UID: "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.355068 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-scripts" (OuterVolumeSpecName: "scripts") pod "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" (UID: "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.380110 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-9c9599df6-zd8gh" podUID="f578a3d5-f846-44b6-bc2b-3d98188dffff" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.386894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" (UID: "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.422400 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-config-data" (OuterVolumeSpecName: "config-data") pod "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" (UID: "9d6b7cfb-7b59-4f0b-b36c-94dae19e512d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.432427 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.432794 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.432825 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.432837 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvt6r\" (UniqueName: \"kubernetes.io/projected/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-kube-api-access-gvt6r\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.432849 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.432858 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.658046 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-ng8zs" event={"ID":"9d6b7cfb-7b59-4f0b-b36c-94dae19e512d","Type":"ContainerDied","Data":"167037b719d8993e64892ba22bab0fb6af36e113f3882870ea0f70d0027d3154"} Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.658362 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="167037b719d8993e64892ba22bab0fb6af36e113f3882870ea0f70d0027d3154" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.658269 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-ng8zs" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.904866 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:09 crc kubenswrapper[4775]: E1126 06:42:09.905861 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" containerName="cinder-db-sync" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.914259 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" containerName="cinder-db-sync" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.914820 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" containerName="cinder-db-sync" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.916193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.923836 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.924124 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.924445 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-dkl9z" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.924589 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.933491 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.943960 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e769a62a-fd58-480a-ad68-010ef2490e7b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.944005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65kf2\" (UniqueName: \"kubernetes.io/projected/e769a62a-fd58-480a-ad68-010ef2490e7b-kube-api-access-65kf2\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.944104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.944135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.944188 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-scripts\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.944225 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:09 crc kubenswrapper[4775]: I1126 06:42:09.958402 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-4xnx9"] Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.006545 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-bchpj"] Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.008108 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.023769 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-bchpj"] Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048683 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048729 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e769a62a-fd58-480a-ad68-010ef2490e7b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048767 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65kf2\" (UniqueName: \"kubernetes.io/projected/e769a62a-fd58-480a-ad68-010ef2490e7b-kube-api-access-65kf2\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048886 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-scripts\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048908 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048931 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7mgz\" (UniqueName: \"kubernetes.io/projected/4770209f-010d-41f0-b211-1bab5637242c-kube-api-access-r7mgz\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.048957 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-config\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.049055 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e769a62a-fd58-480a-ad68-010ef2490e7b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.057460 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.065182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.074218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-scripts\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.091076 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.096153 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65kf2\" (UniqueName: \"kubernetes.io/projected/e769a62a-fd58-480a-ad68-010ef2490e7b-kube-api-access-65kf2\") pod \"cinder-scheduler-0\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.150960 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7mgz\" (UniqueName: \"kubernetes.io/projected/4770209f-010d-41f0-b211-1bab5637242c-kube-api-access-r7mgz\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.151040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-config\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.151099 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.151138 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.151181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.151221 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.152604 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-svc\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.153484 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-config\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.154332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.155007 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.155192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.173961 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7mgz\" (UniqueName: \"kubernetes.io/projected/4770209f-010d-41f0-b211-1bab5637242c-kube-api-access-r7mgz\") pod \"dnsmasq-dns-5784cf869f-bchpj\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.184628 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.230351 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.232768 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.235815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.254957 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-scripts\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.255046 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087188ad-5ba1-46eb-96af-294e39626e0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.255079 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.255118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f9l7\" (UniqueName: \"kubernetes.io/projected/087188ad-5ba1-46eb-96af-294e39626e0b-kube-api-access-6f9l7\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.255156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087188ad-5ba1-46eb-96af-294e39626e0b-logs\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.255222 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.255256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.262548 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.267273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-scripts\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357537 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087188ad-5ba1-46eb-96af-294e39626e0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357571 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357610 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f9l7\" (UniqueName: \"kubernetes.io/projected/087188ad-5ba1-46eb-96af-294e39626e0b-kube-api-access-6f9l7\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087188ad-5ba1-46eb-96af-294e39626e0b-logs\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357707 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.357752 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.363124 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087188ad-5ba1-46eb-96af-294e39626e0b-logs\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.364127 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087188ad-5ba1-46eb-96af-294e39626e0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.366650 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.381376 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.397125 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-scripts\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.399576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.401260 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f9l7\" (UniqueName: \"kubernetes.io/projected/087188ad-5ba1-46eb-96af-294e39626e0b-kube-api-access-6f9l7\") pod \"cinder-api-0\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " pod="openstack/cinder-api-0" Nov 26 06:42:10 crc kubenswrapper[4775]: I1126 06:42:10.603181 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.207913 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.272749 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-config\") pod \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.273016 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-sb\") pod \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.273462 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-swift-storage-0\") pod \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.273488 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-nb\") pod \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.273524 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-svc\") pod \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.273584 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6gs9\" (UniqueName: \"kubernetes.io/projected/fec155eb-aac6-4e4e-a8c7-1487a06671e3-kube-api-access-x6gs9\") pod \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\" (UID: \"fec155eb-aac6-4e4e-a8c7-1487a06671e3\") " Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.279580 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fec155eb-aac6-4e4e-a8c7-1487a06671e3-kube-api-access-x6gs9" (OuterVolumeSpecName: "kube-api-access-x6gs9") pod "fec155eb-aac6-4e4e-a8c7-1487a06671e3" (UID: "fec155eb-aac6-4e4e-a8c7-1487a06671e3"). InnerVolumeSpecName "kube-api-access-x6gs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.334269 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fec155eb-aac6-4e4e-a8c7-1487a06671e3" (UID: "fec155eb-aac6-4e4e-a8c7-1487a06671e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.345420 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fec155eb-aac6-4e4e-a8c7-1487a06671e3" (UID: "fec155eb-aac6-4e4e-a8c7-1487a06671e3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.345800 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fec155eb-aac6-4e4e-a8c7-1487a06671e3" (UID: "fec155eb-aac6-4e4e-a8c7-1487a06671e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.349245 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fec155eb-aac6-4e4e-a8c7-1487a06671e3" (UID: "fec155eb-aac6-4e4e-a8c7-1487a06671e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.375384 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.375411 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.375420 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.375428 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.375437 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6gs9\" (UniqueName: \"kubernetes.io/projected/fec155eb-aac6-4e4e-a8c7-1487a06671e3-kube-api-access-x6gs9\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.392832 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-config" (OuterVolumeSpecName: "config") pod "fec155eb-aac6-4e4e-a8c7-1487a06671e3" (UID: "fec155eb-aac6-4e4e-a8c7-1487a06671e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.476478 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec155eb-aac6-4e4e-a8c7-1487a06671e3-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.683991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" event={"ID":"fec155eb-aac6-4e4e-a8c7-1487a06671e3","Type":"ContainerDied","Data":"8ee82aecb50376f57c5577206da9331339cde15f1db337daa747b6fd8181183d"} Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.684050 4775 scope.go:117] "RemoveContainer" containerID="0aa3eaaabe805177979bb57b4dd85eedb48e6cd8e766aa9381e386554ac49ff8" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.684174 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.730528 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-5twjc"] Nov 26 06:42:11 crc kubenswrapper[4775]: I1126 06:42:11.741247 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-5twjc"] Nov 26 06:42:12 crc kubenswrapper[4775]: I1126 06:42:12.271974 4775 scope.go:117] "RemoveContainer" containerID="498df95d523b87b0f6e1f4d8f7dd33310c768ef1d0e2de356d0814794f78fec6" Nov 26 06:42:12 crc kubenswrapper[4775]: I1126 06:42:12.710690 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fd8c567d4-z5vr4"] Nov 26 06:42:12 crc kubenswrapper[4775]: W1126 06:42:12.726896 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1db6f2a6_c3be_4317_a6ce_802242830778.slice/crio-d3fa21220940efe4faa0c8590fb66409d133e3a16974c65c46b960e8c1cb6b7f WatchSource:0}: Error finding container d3fa21220940efe4faa0c8590fb66409d133e3a16974c65c46b960e8c1cb6b7f: Status 404 returned error can't find the container with id d3fa21220940efe4faa0c8590fb66409d133e3a16974c65c46b960e8c1cb6b7f Nov 26 06:42:12 crc kubenswrapper[4775]: I1126 06:42:12.748274 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-4xnx9"] Nov 26 06:42:12 crc kubenswrapper[4775]: I1126 06:42:12.797848 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:12 crc kubenswrapper[4775]: W1126 06:42:12.830758 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac26de53_1eff_4374_9983_dc657215bf63.slice/crio-f9c5a9ab773997a1d35e020478f8ef6a50612e809c39a4f22d2c260c4cf9bd13 WatchSource:0}: Error finding container f9c5a9ab773997a1d35e020478f8ef6a50612e809c39a4f22d2c260c4cf9bd13: Status 404 returned error can't find the container with id f9c5a9ab773997a1d35e020478f8ef6a50612e809c39a4f22d2c260c4cf9bd13 Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.087772 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cd45b998d-gjx5d"] Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.125757 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:13 crc kubenswrapper[4775]: E1126 06:42:13.223479 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.346766 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" path="/var/lib/kubelet/pods/fec155eb-aac6-4e4e-a8c7-1487a06671e3/volumes" Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.372418 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.405365 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-bchpj"] Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.422250 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dfb875fdb-zr8j4"] Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.434046 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5bdbc86c75-sgb2x"] Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.450245 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-76d59b886b-lsps8"] Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.726880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cd45b998d-gjx5d" event={"ID":"bc391c06-ad32-4c18-9670-481f7614e1a3","Type":"ContainerStarted","Data":"ad49a545d4e019a644571f4fe4a527be364d72acd7cbc42af43f26a90230b22c"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.727258 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cd45b998d-gjx5d" event={"ID":"bc391c06-ad32-4c18-9670-481f7614e1a3","Type":"ContainerStarted","Data":"ad4fbcb8f2b15c7e8813ab23172841d3fd3211fcafb45669befe7e037bedfe88"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.728858 4775 generic.go:334] "Generic (PLEG): container finished" podID="ac26de53-1eff-4374-9983-dc657215bf63" containerID="15480a2a67a406b0a9bf638e0dbc35b09afce150fa721fddd0020cca09184485" exitCode=0 Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.728957 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" event={"ID":"ac26de53-1eff-4374-9983-dc657215bf63","Type":"ContainerDied","Data":"15480a2a67a406b0a9bf638e0dbc35b09afce150fa721fddd0020cca09184485"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.728987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" event={"ID":"ac26de53-1eff-4374-9983-dc657215bf63","Type":"ContainerStarted","Data":"f9c5a9ab773997a1d35e020478f8ef6a50612e809c39a4f22d2c260c4cf9bd13"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.731174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bdbc86c75-sgb2x" event={"ID":"2b958489-5435-42a4-bff8-577a26c717c0","Type":"ContainerStarted","Data":"ceaa341f6d8a127766b78e2a2bc6c048da06f51d368edd9d4ecc555eb018cd1e"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.732924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" event={"ID":"4770209f-010d-41f0-b211-1bab5637242c","Type":"ContainerStarted","Data":"fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.732998 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" event={"ID":"4770209f-010d-41f0-b211-1bab5637242c","Type":"ContainerStarted","Data":"c88b8386087a6b47851e531ad9fed5c0093ca501b4420811d50cc448aeb5f1ad"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.735112 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfb875fdb-zr8j4" event={"ID":"373e990b-3bca-4e7b-8543-840f86d0f7b5","Type":"ContainerStarted","Data":"b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.735146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfb875fdb-zr8j4" event={"ID":"373e990b-3bca-4e7b-8543-840f86d0f7b5","Type":"ContainerStarted","Data":"83c5af063d3652065d0642a28068f10b3c7a9f954ef49126c1ed44dda3121558"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.740680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerStarted","Data":"0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.740839 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.740844 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="ceilometer-notification-agent" containerID="cri-o://ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90" gracePeriod=30 Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.740867 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="sg-core" containerID="cri-o://057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23" gracePeriod=30 Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.740881 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="proxy-httpd" containerID="cri-o://0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40" gracePeriod=30 Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.743533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"087188ad-5ba1-46eb-96af-294e39626e0b","Type":"ContainerStarted","Data":"2e42355f3890dde8a4967faebb2574be62da41048bc006f070dfbceb00c2c31a"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.764537 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" event={"ID":"fa465504-4f8f-4c22-8f6b-3bb834f303c2","Type":"ContainerStarted","Data":"d809d0d4b1c7492d7c5471084de2db9bfc7cc4fec2510564f1014ef4430f8b47"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.770103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fd8c567d4-z5vr4" event={"ID":"1db6f2a6-c3be-4317-a6ce-802242830778","Type":"ContainerStarted","Data":"80acc6780c794b3cacef9175c926d62331f168feadd4bc1cbb46a1bac8daf2e1"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.770146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fd8c567d4-z5vr4" event={"ID":"1db6f2a6-c3be-4317-a6ce-802242830778","Type":"ContainerStarted","Data":"c31bd801ce2b42806c72c1c994105a60b64cc70acc17fff342bdacd8544ecf24"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.770158 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fd8c567d4-z5vr4" event={"ID":"1db6f2a6-c3be-4317-a6ce-802242830778","Type":"ContainerStarted","Data":"d3fa21220940efe4faa0c8590fb66409d133e3a16974c65c46b960e8c1cb6b7f"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.770511 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.770643 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.779923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e769a62a-fd58-480a-ad68-010ef2490e7b","Type":"ContainerStarted","Data":"b032e0219c0addf1301050b1bf68ecfd1be50850d2dc05bcdf97ca9a8b29179b"} Nov 26 06:42:13 crc kubenswrapper[4775]: I1126 06:42:13.809336 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5fd8c567d4-z5vr4" podStartSLOduration=5.8093139350000005 podStartE2EDuration="5.809313935s" podCreationTimestamp="2025-11-26 06:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:13.796958994 +0000 UTC m=+1017.158262956" watchObservedRunningTime="2025-11-26 06:42:13.809313935 +0000 UTC m=+1017.170617887" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.103623 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.257420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-config\") pod \"ac26de53-1eff-4374-9983-dc657215bf63\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.257526 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-svc\") pod \"ac26de53-1eff-4374-9983-dc657215bf63\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.257551 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-sb\") pod \"ac26de53-1eff-4374-9983-dc657215bf63\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.257587 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n48wb\" (UniqueName: \"kubernetes.io/projected/ac26de53-1eff-4374-9983-dc657215bf63-kube-api-access-n48wb\") pod \"ac26de53-1eff-4374-9983-dc657215bf63\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.257653 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-nb\") pod \"ac26de53-1eff-4374-9983-dc657215bf63\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.257748 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-swift-storage-0\") pod \"ac26de53-1eff-4374-9983-dc657215bf63\" (UID: \"ac26de53-1eff-4374-9983-dc657215bf63\") " Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.266115 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac26de53-1eff-4374-9983-dc657215bf63-kube-api-access-n48wb" (OuterVolumeSpecName: "kube-api-access-n48wb") pod "ac26de53-1eff-4374-9983-dc657215bf63" (UID: "ac26de53-1eff-4374-9983-dc657215bf63"). InnerVolumeSpecName "kube-api-access-n48wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.305662 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ac26de53-1eff-4374-9983-dc657215bf63" (UID: "ac26de53-1eff-4374-9983-dc657215bf63"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.309103 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac26de53-1eff-4374-9983-dc657215bf63" (UID: "ac26de53-1eff-4374-9983-dc657215bf63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.337957 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac26de53-1eff-4374-9983-dc657215bf63" (UID: "ac26de53-1eff-4374-9983-dc657215bf63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.342689 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-config" (OuterVolumeSpecName: "config") pod "ac26de53-1eff-4374-9983-dc657215bf63" (UID: "ac26de53-1eff-4374-9983-dc657215bf63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.349598 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ac26de53-1eff-4374-9983-dc657215bf63" (UID: "ac26de53-1eff-4374-9983-dc657215bf63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.360928 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.361138 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.361158 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.361170 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n48wb\" (UniqueName: \"kubernetes.io/projected/ac26de53-1eff-4374-9983-dc657215bf63-kube-api-access-n48wb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.361179 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.361188 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac26de53-1eff-4374-9983-dc657215bf63-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.405660 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84b966f6c9-5twjc" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.151:5353: i/o timeout" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.789942 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfb875fdb-zr8j4" event={"ID":"373e990b-3bca-4e7b-8543-840f86d0f7b5","Type":"ContainerStarted","Data":"5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.790032 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.792291 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.792289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-4xnx9" event={"ID":"ac26de53-1eff-4374-9983-dc657215bf63","Type":"ContainerDied","Data":"f9c5a9ab773997a1d35e020478f8ef6a50612e809c39a4f22d2c260c4cf9bd13"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.792492 4775 scope.go:117] "RemoveContainer" containerID="15480a2a67a406b0a9bf638e0dbc35b09afce150fa721fddd0020cca09184485" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.809108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerDied","Data":"0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.809010 4775 generic.go:334] "Generic (PLEG): container finished" podID="a14e76fa-5315-4564-9430-9231a8e31a01" containerID="0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40" exitCode=0 Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.814533 4775 generic.go:334] "Generic (PLEG): container finished" podID="a14e76fa-5315-4564-9430-9231a8e31a01" containerID="057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23" exitCode=2 Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.814661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerDied","Data":"057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.817835 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podStartSLOduration=9.817816802 podStartE2EDuration="9.817816802s" podCreationTimestamp="2025-11-26 06:42:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:14.807546725 +0000 UTC m=+1018.168850687" watchObservedRunningTime="2025-11-26 06:42:14.817816802 +0000 UTC m=+1018.179120754" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.820780 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"087188ad-5ba1-46eb-96af-294e39626e0b","Type":"ContainerStarted","Data":"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.822929 4775 generic.go:334] "Generic (PLEG): container finished" podID="4770209f-010d-41f0-b211-1bab5637242c" containerID="fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0" exitCode=0 Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.822991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" event={"ID":"4770209f-010d-41f0-b211-1bab5637242c","Type":"ContainerDied","Data":"fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.830777 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cd45b998d-gjx5d" event={"ID":"bc391c06-ad32-4c18-9670-481f7614e1a3","Type":"ContainerStarted","Data":"d513fbb1056c7f1a6d2a98a6075dd1638d9039f1c1b5b61b3a8df2f9b9d04289"} Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.830829 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.830863 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.878964 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-4xnx9"] Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.891919 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-4xnx9"] Nov 26 06:42:14 crc kubenswrapper[4775]: I1126 06:42:14.907096 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5cd45b998d-gjx5d" podStartSLOduration=10.907074723000001 podStartE2EDuration="10.907074723s" podCreationTimestamp="2025-11-26 06:42:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:14.894846115 +0000 UTC m=+1018.256150087" watchObservedRunningTime="2025-11-26 06:42:14.907074723 +0000 UTC m=+1018.268378675" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.353273 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac26de53-1eff-4374-9983-dc657215bf63" path="/var/lib/kubelet/pods/ac26de53-1eff-4374-9983-dc657215bf63/volumes" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.494665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.808177 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.851111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" event={"ID":"fa465504-4f8f-4c22-8f6b-3bb834f303c2","Type":"ContainerStarted","Data":"cd46015543b1db0655ab4d2819f710996cb92e7c17945fffec11f33a1180a9e7"} Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.853296 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bdbc86c75-sgb2x" event={"ID":"2b958489-5435-42a4-bff8-577a26c717c0","Type":"ContainerStarted","Data":"a5943f0d1e60b4d532cc8897c10a0235cf2a8ded2ce49cfb556bb8a377c06474"} Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.859019 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" event={"ID":"4770209f-010d-41f0-b211-1bab5637242c","Type":"ContainerStarted","Data":"b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585"} Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.859124 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.871207 4775 generic.go:334] "Generic (PLEG): container finished" podID="a14e76fa-5315-4564-9430-9231a8e31a01" containerID="ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90" exitCode=0 Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.871328 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.871580 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerDied","Data":"ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90"} Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.871631 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a14e76fa-5315-4564-9430-9231a8e31a01","Type":"ContainerDied","Data":"966d923c6267562e6b22c451aa680e25509daf2a420c16f716f81e13ba366f8f"} Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.871649 4775 scope.go:117] "RemoveContainer" containerID="0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.874980 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"087188ad-5ba1-46eb-96af-294e39626e0b","Type":"ContainerStarted","Data":"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513"} Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.875108 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api-log" containerID="cri-o://d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98" gracePeriod=30 Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.875111 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api" containerID="cri-o://90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513" gracePeriod=30 Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.888639 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" podStartSLOduration=6.888618269 podStartE2EDuration="6.888618269s" podCreationTimestamp="2025-11-26 06:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:15.886937285 +0000 UTC m=+1019.248241257" watchObservedRunningTime="2025-11-26 06:42:15.888618269 +0000 UTC m=+1019.249922231" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.915881 4775 scope.go:117] "RemoveContainer" containerID="057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.922553 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.922536061 podStartE2EDuration="5.922536061s" podCreationTimestamp="2025-11-26 06:42:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:15.916096163 +0000 UTC m=+1019.277400115" watchObservedRunningTime="2025-11-26 06:42:15.922536061 +0000 UTC m=+1019.283840003" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.944883 4775 scope.go:117] "RemoveContainer" containerID="ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.968640 4775 scope.go:117] "RemoveContainer" containerID="0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40" Nov 26 06:42:15 crc kubenswrapper[4775]: E1126 06:42:15.969355 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40\": container with ID starting with 0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40 not found: ID does not exist" containerID="0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.969409 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40"} err="failed to get container status \"0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40\": rpc error: code = NotFound desc = could not find container \"0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40\": container with ID starting with 0d8e616261e3905422574357c699c1994d39f1adcca0079d3b7bdf3043144f40 not found: ID does not exist" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.969444 4775 scope.go:117] "RemoveContainer" containerID="057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23" Nov 26 06:42:15 crc kubenswrapper[4775]: E1126 06:42:15.969995 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23\": container with ID starting with 057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23 not found: ID does not exist" containerID="057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.970045 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23"} err="failed to get container status \"057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23\": rpc error: code = NotFound desc = could not find container \"057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23\": container with ID starting with 057ca8996c8c295c581811d67c85a81fba6b6f0c0d6f9e74248ed66373785f23 not found: ID does not exist" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.970078 4775 scope.go:117] "RemoveContainer" containerID="ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90" Nov 26 06:42:15 crc kubenswrapper[4775]: E1126 06:42:15.970462 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90\": container with ID starting with ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90 not found: ID does not exist" containerID="ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90" Nov 26 06:42:15 crc kubenswrapper[4775]: I1126 06:42:15.970506 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90"} err="failed to get container status \"ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90\": rpc error: code = NotFound desc = could not find container \"ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90\": container with ID starting with ed969fb33bdb74630feabfe3a3d04a9ae01f902821430fb0f83ec6362c112d90 not found: ID does not exist" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000244 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-log-httpd\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000289 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-combined-ca-bundle\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000317 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5wwk\" (UniqueName: \"kubernetes.io/projected/a14e76fa-5315-4564-9430-9231a8e31a01-kube-api-access-g5wwk\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000359 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-run-httpd\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000414 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-config-data\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000583 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-scripts\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.000646 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-sg-core-conf-yaml\") pod \"a14e76fa-5315-4564-9430-9231a8e31a01\" (UID: \"a14e76fa-5315-4564-9430-9231a8e31a01\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.005053 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.007869 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a14e76fa-5315-4564-9430-9231a8e31a01-kube-api-access-g5wwk" (OuterVolumeSpecName: "kube-api-access-g5wwk") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "kube-api-access-g5wwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.009150 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.014926 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-scripts" (OuterVolumeSpecName: "scripts") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.056119 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.082064 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.093527 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-config-data" (OuterVolumeSpecName: "config-data") pod "a14e76fa-5315-4564-9430-9231a8e31a01" (UID: "a14e76fa-5315-4564-9430-9231a8e31a01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103394 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103440 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103458 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103472 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103484 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5wwk\" (UniqueName: \"kubernetes.io/projected/a14e76fa-5315-4564-9430-9231a8e31a01-kube-api-access-g5wwk\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103496 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a14e76fa-5315-4564-9430-9231a8e31a01-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.103508 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a14e76fa-5315-4564-9430-9231a8e31a01-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.266034 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.278689 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.289777 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:16 crc kubenswrapper[4775]: E1126 06:42:16.290243 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="dnsmasq-dns" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290256 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="dnsmasq-dns" Nov 26 06:42:16 crc kubenswrapper[4775]: E1126 06:42:16.290278 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac26de53-1eff-4374-9983-dc657215bf63" containerName="init" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290285 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac26de53-1eff-4374-9983-dc657215bf63" containerName="init" Nov 26 06:42:16 crc kubenswrapper[4775]: E1126 06:42:16.290295 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="ceilometer-notification-agent" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290302 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="ceilometer-notification-agent" Nov 26 06:42:16 crc kubenswrapper[4775]: E1126 06:42:16.290318 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="init" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290324 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="init" Nov 26 06:42:16 crc kubenswrapper[4775]: E1126 06:42:16.290332 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="sg-core" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290338 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="sg-core" Nov 26 06:42:16 crc kubenswrapper[4775]: E1126 06:42:16.290347 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="proxy-httpd" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290353 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="proxy-httpd" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290535 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="proxy-httpd" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290556 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="ceilometer-notification-agent" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290568 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fec155eb-aac6-4e4e-a8c7-1487a06671e3" containerName="dnsmasq-dns" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290585 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac26de53-1eff-4374-9983-dc657215bf63" containerName="init" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.290598 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" containerName="sg-core" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.292285 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.293469 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.297556 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.297765 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315380 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-log-httpd\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315553 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315614 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpqs5\" (UniqueName: \"kubernetes.io/projected/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-kube-api-access-jpqs5\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315698 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-config-data\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-scripts\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.315882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-run-httpd\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.418393 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-log-httpd\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.418860 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.418891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.418913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpqs5\" (UniqueName: \"kubernetes.io/projected/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-kube-api-access-jpqs5\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.419004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-config-data\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.419086 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-scripts\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.419151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-run-httpd\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.419770 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-run-httpd\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.420156 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-log-httpd\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.428436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-config-data\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.429246 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.432481 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-scripts\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.432930 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.441839 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpqs5\" (UniqueName: \"kubernetes.io/projected/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-kube-api-access-jpqs5\") pod \"ceilometer-0\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.471073 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539276 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data-custom\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539340 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539378 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087188ad-5ba1-46eb-96af-294e39626e0b-etc-machine-id\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539488 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-scripts\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539538 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f9l7\" (UniqueName: \"kubernetes.io/projected/087188ad-5ba1-46eb-96af-294e39626e0b-kube-api-access-6f9l7\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539619 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-combined-ca-bundle\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.539679 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087188ad-5ba1-46eb-96af-294e39626e0b-logs\") pod \"087188ad-5ba1-46eb-96af-294e39626e0b\" (UID: \"087188ad-5ba1-46eb-96af-294e39626e0b\") " Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.540461 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/087188ad-5ba1-46eb-96af-294e39626e0b-logs" (OuterVolumeSpecName: "logs") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.541852 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087188ad-5ba1-46eb-96af-294e39626e0b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.550098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.554911 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-scripts" (OuterVolumeSpecName: "scripts") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.556004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087188ad-5ba1-46eb-96af-294e39626e0b-kube-api-access-6f9l7" (OuterVolumeSpecName: "kube-api-access-6f9l7") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "kube-api-access-6f9l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.621805 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.631360 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.634908 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data" (OuterVolumeSpecName: "config-data") pod "087188ad-5ba1-46eb-96af-294e39626e0b" (UID: "087188ad-5ba1-46eb-96af-294e39626e0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657625 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657656 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/087188ad-5ba1-46eb-96af-294e39626e0b-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657665 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657674 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657684 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/087188ad-5ba1-46eb-96af-294e39626e0b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657692 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/087188ad-5ba1-46eb-96af-294e39626e0b-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.657700 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f9l7\" (UniqueName: \"kubernetes.io/projected/087188ad-5ba1-46eb-96af-294e39626e0b-kube-api-access-6f9l7\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.891953 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" event={"ID":"fa465504-4f8f-4c22-8f6b-3bb834f303c2","Type":"ContainerStarted","Data":"85374c20ab7075516ad7fc770388949141a64a6906c556143aa543037e1104f3"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.894121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5bdbc86c75-sgb2x" event={"ID":"2b958489-5435-42a4-bff8-577a26c717c0","Type":"ContainerStarted","Data":"96a03b7ccfd6d352a63b55b2909c5756d5cfd0421e03de43193e6e6ec3e23263"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.900281 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e769a62a-fd58-480a-ad68-010ef2490e7b","Type":"ContainerStarted","Data":"c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.900326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e769a62a-fd58-480a-ad68-010ef2490e7b","Type":"ContainerStarted","Data":"6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.908003 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-76d59b886b-lsps8" podStartSLOduration=10.995183874 podStartE2EDuration="12.907986269s" podCreationTimestamp="2025-11-26 06:42:04 +0000 UTC" firstStartedPulling="2025-11-26 06:42:13.506857033 +0000 UTC m=+1016.868160985" lastFinishedPulling="2025-11-26 06:42:15.419659428 +0000 UTC m=+1018.780963380" observedRunningTime="2025-11-26 06:42:16.907337992 +0000 UTC m=+1020.268641934" watchObservedRunningTime="2025-11-26 06:42:16.907986269 +0000 UTC m=+1020.269290221" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914273 4775 generic.go:334] "Generic (PLEG): container finished" podID="087188ad-5ba1-46eb-96af-294e39626e0b" containerID="90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513" exitCode=0 Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914321 4775 generic.go:334] "Generic (PLEG): container finished" podID="087188ad-5ba1-46eb-96af-294e39626e0b" containerID="d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98" exitCode=143 Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914789 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"087188ad-5ba1-46eb-96af-294e39626e0b","Type":"ContainerDied","Data":"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914832 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"087188ad-5ba1-46eb-96af-294e39626e0b","Type":"ContainerDied","Data":"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"087188ad-5ba1-46eb-96af-294e39626e0b","Type":"ContainerDied","Data":"2e42355f3890dde8a4967faebb2574be62da41048bc006f070dfbceb00c2c31a"} Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914857 4775 scope.go:117] "RemoveContainer" containerID="90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.914973 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.934345 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5bdbc86c75-sgb2x" podStartSLOduration=10.990855652 podStartE2EDuration="12.934321423s" podCreationTimestamp="2025-11-26 06:42:04 +0000 UTC" firstStartedPulling="2025-11-26 06:42:13.47786122 +0000 UTC m=+1016.839165172" lastFinishedPulling="2025-11-26 06:42:15.421326991 +0000 UTC m=+1018.782630943" observedRunningTime="2025-11-26 06:42:16.923203104 +0000 UTC m=+1020.284507076" watchObservedRunningTime="2025-11-26 06:42:16.934321423 +0000 UTC m=+1020.295625375" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.959842 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.875576921 podStartE2EDuration="7.959821106s" podCreationTimestamp="2025-11-26 06:42:09 +0000 UTC" firstStartedPulling="2025-11-26 06:42:13.457912411 +0000 UTC m=+1016.819216363" lastFinishedPulling="2025-11-26 06:42:14.542156596 +0000 UTC m=+1017.903460548" observedRunningTime="2025-11-26 06:42:16.954269422 +0000 UTC m=+1020.315573394" watchObservedRunningTime="2025-11-26 06:42:16.959821106 +0000 UTC m=+1020.321125058" Nov 26 06:42:16 crc kubenswrapper[4775]: I1126 06:42:16.993065 4775 scope.go:117] "RemoveContainer" containerID="d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.031828 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.036898 4775 scope.go:117] "RemoveContainer" containerID="90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513" Nov 26 06:42:17 crc kubenswrapper[4775]: E1126 06:42:17.040354 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513\": container with ID starting with 90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513 not found: ID does not exist" containerID="90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.040402 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513"} err="failed to get container status \"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513\": rpc error: code = NotFound desc = could not find container \"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513\": container with ID starting with 90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513 not found: ID does not exist" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.040427 4775 scope.go:117] "RemoveContainer" containerID="d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98" Nov 26 06:42:17 crc kubenswrapper[4775]: E1126 06:42:17.057285 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98\": container with ID starting with d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98 not found: ID does not exist" containerID="d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.057377 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98"} err="failed to get container status \"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98\": rpc error: code = NotFound desc = could not find container \"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98\": container with ID starting with d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98 not found: ID does not exist" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.057521 4775 scope.go:117] "RemoveContainer" containerID="90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.057655 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.058016 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513"} err="failed to get container status \"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513\": rpc error: code = NotFound desc = could not find container \"90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513\": container with ID starting with 90afe64d07f25726f05649b47a8d7bba07893b412c173b1c956bbfceb736e513 not found: ID does not exist" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.058055 4775 scope.go:117] "RemoveContainer" containerID="d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.059064 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98"} err="failed to get container status \"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98\": rpc error: code = NotFound desc = could not find container \"d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98\": container with ID starting with d3ecfd3d2ba481f7d63c131f50b331a85722e17be03a6fcbb72e6aec9f0aad98 not found: ID does not exist" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.067130 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:17 crc kubenswrapper[4775]: E1126 06:42:17.067627 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.067645 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api" Nov 26 06:42:17 crc kubenswrapper[4775]: E1126 06:42:17.067680 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api-log" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.067688 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api-log" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.067923 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api-log" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.067953 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" containerName="cinder-api" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.068915 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.072125 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.072177 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.072583 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.080702 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.121326 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.173949 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.173995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174022 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-config-data-custom\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174037 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-scripts\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b9daeb8-621b-4557-9757-6e8b9c430339-logs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174139 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174162 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b9daeb8-621b-4557-9757-6e8b9c430339-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pqpg\" (UniqueName: \"kubernetes.io/projected/4b9daeb8-621b-4557-9757-6e8b9c430339-kube-api-access-7pqpg\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.174212 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-config-data\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.275793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-config-data\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.275873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.275905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.275941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-config-data-custom\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.275961 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-scripts\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.276069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b9daeb8-621b-4557-9757-6e8b9c430339-logs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.276114 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.276147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b9daeb8-621b-4557-9757-6e8b9c430339-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.276173 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pqpg\" (UniqueName: \"kubernetes.io/projected/4b9daeb8-621b-4557-9757-6e8b9c430339-kube-api-access-7pqpg\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.277201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b9daeb8-621b-4557-9757-6e8b9c430339-logs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.278624 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b9daeb8-621b-4557-9757-6e8b9c430339-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.279051 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.279131 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.283219 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.284157 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-scripts\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.289697 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.292406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-config-data-custom\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.293538 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.293804 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-config-data\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.295226 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b9daeb8-621b-4557-9757-6e8b9c430339-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.297039 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pqpg\" (UniqueName: \"kubernetes.io/projected/4b9daeb8-621b-4557-9757-6e8b9c430339-kube-api-access-7pqpg\") pod \"cinder-api-0\" (UID: \"4b9daeb8-621b-4557-9757-6e8b9c430339\") " pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.347257 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087188ad-5ba1-46eb-96af-294e39626e0b" path="/var/lib/kubelet/pods/087188ad-5ba1-46eb-96af-294e39626e0b/volumes" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.348348 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a14e76fa-5315-4564-9430-9231a8e31a01" path="/var/lib/kubelet/pods/a14e76fa-5315-4564-9430-9231a8e31a01/volumes" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.396327 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.882112 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 26 06:42:17 crc kubenswrapper[4775]: W1126 06:42:17.888588 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b9daeb8_621b_4557_9757_6e8b9c430339.slice/crio-9c55a1b67c2697b9197ff3b1440e35a64bbdad8af86bf25aacd100027d5a2f91 WatchSource:0}: Error finding container 9c55a1b67c2697b9197ff3b1440e35a64bbdad8af86bf25aacd100027d5a2f91: Status 404 returned error can't find the container with id 9c55a1b67c2697b9197ff3b1440e35a64bbdad8af86bf25aacd100027d5a2f91 Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.938032 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4b9daeb8-621b-4557-9757-6e8b9c430339","Type":"ContainerStarted","Data":"9c55a1b67c2697b9197ff3b1440e35a64bbdad8af86bf25aacd100027d5a2f91"} Nov 26 06:42:17 crc kubenswrapper[4775]: I1126 06:42:17.940341 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerStarted","Data":"8b00f2498c4b78cf4c91aca8036a14b0752d66bce79c755506852c7bfe5ec650"} Nov 26 06:42:18 crc kubenswrapper[4775]: I1126 06:42:18.950777 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerStarted","Data":"13559158189cc3bd1149718ea6f4241e0a4c9fa4bfc7871e682ed45f85c0db63"} Nov 26 06:42:18 crc kubenswrapper[4775]: I1126 06:42:18.951272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerStarted","Data":"c436d72e873c9a9792d542db4e5fea82b2eb902503fa297c01bc68556c2ddb37"} Nov 26 06:42:18 crc kubenswrapper[4775]: I1126 06:42:18.954853 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4b9daeb8-621b-4557-9757-6e8b9c430339","Type":"ContainerStarted","Data":"7cdd71ba1ade5953a70c37078efeaac959ae28ab458cdcd02a25915566e968cb"} Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.980502 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4b9daeb8-621b-4557-9757-6e8b9c430339","Type":"ContainerStarted","Data":"fc737659ffe38d0584bee8fe415ff8243fc7771f3f0b8cc1aec17b8fcd811485"} Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.980993 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.982945 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerStarted","Data":"ec61dec025684a387060b9d9a08a3ce4ac282766f418f44300ebb8daf241e508"} Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.985241 4775 generic.go:334] "Generic (PLEG): container finished" podID="10c96169-294e-4273-8e08-92c677dfc01c" containerID="0d33e092249e992f336e7facec4df4d9ffea73cb6fe83a88efd5ff023dbb300f" exitCode=137 Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.985279 4775 generic.go:334] "Generic (PLEG): container finished" podID="10c96169-294e-4273-8e08-92c677dfc01c" containerID="cbb5041fedb73961d9aac8c4248c3cdd509372660fff75d643233a535805de73" exitCode=137 Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.985300 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84989d94bf-7vmkz" event={"ID":"10c96169-294e-4273-8e08-92c677dfc01c","Type":"ContainerDied","Data":"0d33e092249e992f336e7facec4df4d9ffea73cb6fe83a88efd5ff023dbb300f"} Nov 26 06:42:19 crc kubenswrapper[4775]: I1126 06:42:19.985325 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84989d94bf-7vmkz" event={"ID":"10c96169-294e-4273-8e08-92c677dfc01c","Type":"ContainerDied","Data":"cbb5041fedb73961d9aac8c4248c3cdd509372660fff75d643233a535805de73"} Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.008944 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.008927311 podStartE2EDuration="4.008927311s" podCreationTimestamp="2025-11-26 06:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:20.002407011 +0000 UTC m=+1023.363710973" watchObservedRunningTime="2025-11-26 06:42:20.008927311 +0000 UTC m=+1023.370231263" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.057608 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.134404 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-config-data\") pod \"10c96169-294e-4273-8e08-92c677dfc01c\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.134549 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c96169-294e-4273-8e08-92c677dfc01c-logs\") pod \"10c96169-294e-4273-8e08-92c677dfc01c\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.134642 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts\") pod \"10c96169-294e-4273-8e08-92c677dfc01c\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.134671 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rfxs\" (UniqueName: \"kubernetes.io/projected/10c96169-294e-4273-8e08-92c677dfc01c-kube-api-access-6rfxs\") pod \"10c96169-294e-4273-8e08-92c677dfc01c\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.134739 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/10c96169-294e-4273-8e08-92c677dfc01c-horizon-secret-key\") pod \"10c96169-294e-4273-8e08-92c677dfc01c\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.135299 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10c96169-294e-4273-8e08-92c677dfc01c-logs" (OuterVolumeSpecName: "logs") pod "10c96169-294e-4273-8e08-92c677dfc01c" (UID: "10c96169-294e-4273-8e08-92c677dfc01c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.143887 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10c96169-294e-4273-8e08-92c677dfc01c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "10c96169-294e-4273-8e08-92c677dfc01c" (UID: "10c96169-294e-4273-8e08-92c677dfc01c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.161984 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10c96169-294e-4273-8e08-92c677dfc01c-kube-api-access-6rfxs" (OuterVolumeSpecName: "kube-api-access-6rfxs") pod "10c96169-294e-4273-8e08-92c677dfc01c" (UID: "10c96169-294e-4273-8e08-92c677dfc01c"). InnerVolumeSpecName "kube-api-access-6rfxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:20 crc kubenswrapper[4775]: E1126 06:42:20.162046 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts podName:10c96169-294e-4273-8e08-92c677dfc01c nodeName:}" failed. No retries permitted until 2025-11-26 06:42:20.66201418 +0000 UTC m=+1024.023318142 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "scripts" (UniqueName: "kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts") pod "10c96169-294e-4273-8e08-92c677dfc01c" (UID: "10c96169-294e-4273-8e08-92c677dfc01c") : error deleting /var/lib/kubelet/pods/10c96169-294e-4273-8e08-92c677dfc01c/volume-subpaths: remove /var/lib/kubelet/pods/10c96169-294e-4273-8e08-92c677dfc01c/volume-subpaths: no such file or directory Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.162501 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-config-data" (OuterVolumeSpecName: "config-data") pod "10c96169-294e-4273-8e08-92c677dfc01c" (UID: "10c96169-294e-4273-8e08-92c677dfc01c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.187955 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.239968 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-cs4rl"] Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.240205 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="dnsmasq-dns" containerID="cri-o://75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2" gracePeriod=10 Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.241784 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10c96169-294e-4273-8e08-92c677dfc01c-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.241797 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rfxs\" (UniqueName: \"kubernetes.io/projected/10c96169-294e-4273-8e08-92c677dfc01c-kube-api-access-6rfxs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.241807 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/10c96169-294e-4273-8e08-92c677dfc01c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.241817 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.269883 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.701046 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.757048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts\") pod \"10c96169-294e-4273-8e08-92c677dfc01c\" (UID: \"10c96169-294e-4273-8e08-92c677dfc01c\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.765348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts" (OuterVolumeSpecName: "scripts") pod "10c96169-294e-4273-8e08-92c677dfc01c" (UID: "10c96169-294e-4273-8e08-92c677dfc01c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.859654 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10c96169-294e-4273-8e08-92c677dfc01c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.865246 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fd8c567d4-z5vr4" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.878026 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.939448 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5dfb875fdb-zr8j4"] Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.940015 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" containerID="cri-o://b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59" gracePeriod=30 Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.940548 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" containerID="cri-o://5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d" gracePeriod=30 Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.958004 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.958570 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.959043 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.959276 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": EOF" Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.960447 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-sb\") pod \"acbdc4be-0331-410a-8f6c-b87323d59456\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.960525 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-nb\") pod \"acbdc4be-0331-410a-8f6c-b87323d59456\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.960576 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msnr2\" (UniqueName: \"kubernetes.io/projected/acbdc4be-0331-410a-8f6c-b87323d59456-kube-api-access-msnr2\") pod \"acbdc4be-0331-410a-8f6c-b87323d59456\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.960622 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-config\") pod \"acbdc4be-0331-410a-8f6c-b87323d59456\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.960667 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-svc\") pod \"acbdc4be-0331-410a-8f6c-b87323d59456\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.960726 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-swift-storage-0\") pod \"acbdc4be-0331-410a-8f6c-b87323d59456\" (UID: \"acbdc4be-0331-410a-8f6c-b87323d59456\") " Nov 26 06:42:20 crc kubenswrapper[4775]: I1126 06:42:20.970999 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acbdc4be-0331-410a-8f6c-b87323d59456-kube-api-access-msnr2" (OuterVolumeSpecName: "kube-api-access-msnr2") pod "acbdc4be-0331-410a-8f6c-b87323d59456" (UID: "acbdc4be-0331-410a-8f6c-b87323d59456"). InnerVolumeSpecName "kube-api-access-msnr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.019360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84989d94bf-7vmkz" event={"ID":"10c96169-294e-4273-8e08-92c677dfc01c","Type":"ContainerDied","Data":"ca53cabd1ed7c8756b01398673f8a23343b8cc29c51f966fa0f9917fd93a18dd"} Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.019414 4775 scope.go:117] "RemoveContainer" containerID="0d33e092249e992f336e7facec4df4d9ffea73cb6fe83a88efd5ff023dbb300f" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.019528 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84989d94bf-7vmkz" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.053976 4775 generic.go:334] "Generic (PLEG): container finished" podID="acbdc4be-0331-410a-8f6c-b87323d59456" containerID="75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2" exitCode=0 Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.054926 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.055293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" event={"ID":"acbdc4be-0331-410a-8f6c-b87323d59456","Type":"ContainerDied","Data":"75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2"} Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.055315 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" event={"ID":"acbdc4be-0331-410a-8f6c-b87323d59456","Type":"ContainerDied","Data":"1240e7f8f1ab4980af1572bd8249739fa485ff91505dd51633cae09c4e743bf1"} Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.072953 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msnr2\" (UniqueName: \"kubernetes.io/projected/acbdc4be-0331-410a-8f6c-b87323d59456-kube-api-access-msnr2\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.109978 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "acbdc4be-0331-410a-8f6c-b87323d59456" (UID: "acbdc4be-0331-410a-8f6c-b87323d59456"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.119514 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-config" (OuterVolumeSpecName: "config") pod "acbdc4be-0331-410a-8f6c-b87323d59456" (UID: "acbdc4be-0331-410a-8f6c-b87323d59456"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.130164 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "acbdc4be-0331-410a-8f6c-b87323d59456" (UID: "acbdc4be-0331-410a-8f6c-b87323d59456"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.149327 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "acbdc4be-0331-410a-8f6c-b87323d59456" (UID: "acbdc4be-0331-410a-8f6c-b87323d59456"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.157641 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84989d94bf-7vmkz"] Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.166846 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-84989d94bf-7vmkz"] Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.178004 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.178036 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.178045 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.178053 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.228817 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "acbdc4be-0331-410a-8f6c-b87323d59456" (UID: "acbdc4be-0331-410a-8f6c-b87323d59456"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.280760 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/acbdc4be-0331-410a-8f6c-b87323d59456-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.309780 4775 scope.go:117] "RemoveContainer" containerID="cbb5041fedb73961d9aac8c4248c3cdd509372660fff75d643233a535805de73" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.344905 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10c96169-294e-4273-8e08-92c677dfc01c" path="/var/lib/kubelet/pods/10c96169-294e-4273-8e08-92c677dfc01c/volumes" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.370925 4775 scope.go:117] "RemoveContainer" containerID="75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.410877 4775 scope.go:117] "RemoveContainer" containerID="04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.434289 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-cs4rl"] Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.440891 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-cs4rl"] Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.475423 4775 scope.go:117] "RemoveContainer" containerID="75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2" Nov 26 06:42:21 crc kubenswrapper[4775]: E1126 06:42:21.477462 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2\": container with ID starting with 75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2 not found: ID does not exist" containerID="75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.477540 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2"} err="failed to get container status \"75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2\": rpc error: code = NotFound desc = could not find container \"75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2\": container with ID starting with 75a40ea710a759e933b4ad823d2d0a6b8dccea6836a467950f9c9c9e69485ab2 not found: ID does not exist" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.477563 4775 scope.go:117] "RemoveContainer" containerID="04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba" Nov 26 06:42:21 crc kubenswrapper[4775]: E1126 06:42:21.479650 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba\": container with ID starting with 04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba not found: ID does not exist" containerID="04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba" Nov 26 06:42:21 crc kubenswrapper[4775]: I1126 06:42:21.479784 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba"} err="failed to get container status \"04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba\": rpc error: code = NotFound desc = could not find container \"04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba\": container with ID starting with 04dacc8733d1c89db333801d2817ec3f9f360fb1d7712798d2bd9b4c12c84fba not found: ID does not exist" Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.026764 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.081375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerStarted","Data":"8e01992a6ecd9cc660a39d7675118556b27a160f4dd39c30a8a98e06bd04a641"} Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.082500 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.086572 4775 generic.go:334] "Generic (PLEG): container finished" podID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerID="b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59" exitCode=143 Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.086609 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfb875fdb-zr8j4" event={"ID":"373e990b-3bca-4e7b-8543-840f86d0f7b5","Type":"ContainerDied","Data":"b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59"} Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.240993 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:42:22 crc kubenswrapper[4775]: I1126 06:42:22.268793 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.394704738 podStartE2EDuration="6.268772177s" podCreationTimestamp="2025-11-26 06:42:16 +0000 UTC" firstStartedPulling="2025-11-26 06:42:17.126854429 +0000 UTC m=+1020.488158381" lastFinishedPulling="2025-11-26 06:42:21.000921858 +0000 UTC m=+1024.362225820" observedRunningTime="2025-11-26 06:42:22.107423642 +0000 UTC m=+1025.468727604" watchObservedRunningTime="2025-11-26 06:42:22.268772177 +0000 UTC m=+1025.630076129" Nov 26 06:42:23 crc kubenswrapper[4775]: I1126 06:42:23.350337 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" path="/var/lib/kubelet/pods/acbdc4be-0331-410a-8f6c-b87323d59456/volumes" Nov 26 06:42:23 crc kubenswrapper[4775]: I1126 06:42:23.936550 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:42:23 crc kubenswrapper[4775]: I1126 06:42:23.941533 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-9c9599df6-zd8gh" Nov 26 06:42:24 crc kubenswrapper[4775]: I1126 06:42:24.028011 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5886fff4d-zcj7s"] Nov 26 06:42:24 crc kubenswrapper[4775]: I1126 06:42:24.106164 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon-log" containerID="cri-o://aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0" gracePeriod=30 Nov 26 06:42:24 crc kubenswrapper[4775]: I1126 06:42:24.108223 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" containerID="cri-o://4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a" gracePeriod=30 Nov 26 06:42:25 crc kubenswrapper[4775]: I1126 06:42:25.486626 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 26 06:42:25 crc kubenswrapper[4775]: I1126 06:42:25.561578 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:25 crc kubenswrapper[4775]: I1126 06:42:25.691481 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8b5c85b87-cs4rl" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Nov 26 06:42:26 crc kubenswrapper[4775]: I1126 06:42:26.040955 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 06:42:26 crc kubenswrapper[4775]: I1126 06:42:26.040990 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 06:42:26 crc kubenswrapper[4775]: I1126 06:42:26.121494 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="cinder-scheduler" containerID="cri-o://6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5" gracePeriod=30 Nov 26 06:42:26 crc kubenswrapper[4775]: I1126 06:42:26.121555 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="probe" containerID="cri-o://c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a" gracePeriod=30 Nov 26 06:42:26 crc kubenswrapper[4775]: I1126 06:42:26.369684 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:48128->10.217.0.158:9311: read: connection reset by peer" Nov 26 06:42:26 crc kubenswrapper[4775]: I1126 06:42:26.369782 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dfb875fdb-zr8j4" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": read tcp 10.217.0.2:48136->10.217.0.158:9311: read: connection reset by peer" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.102307 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.132230 4775 generic.go:334] "Generic (PLEG): container finished" podID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerID="5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d" exitCode=0 Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.132302 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfb875fdb-zr8j4" event={"ID":"373e990b-3bca-4e7b-8543-840f86d0f7b5","Type":"ContainerDied","Data":"5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d"} Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.132326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dfb875fdb-zr8j4" event={"ID":"373e990b-3bca-4e7b-8543-840f86d0f7b5","Type":"ContainerDied","Data":"83c5af063d3652065d0642a28068f10b3c7a9f954ef49126c1ed44dda3121558"} Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.132342 4775 scope.go:117] "RemoveContainer" containerID="5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.132571 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dfb875fdb-zr8j4" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.136147 4775 generic.go:334] "Generic (PLEG): container finished" podID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerID="c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a" exitCode=0 Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.136180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e769a62a-fd58-480a-ad68-010ef2490e7b","Type":"ContainerDied","Data":"c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a"} Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.197612 4775 scope.go:117] "RemoveContainer" containerID="b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.206262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhwrd\" (UniqueName: \"kubernetes.io/projected/373e990b-3bca-4e7b-8543-840f86d0f7b5-kube-api-access-hhwrd\") pod \"373e990b-3bca-4e7b-8543-840f86d0f7b5\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.206414 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data\") pod \"373e990b-3bca-4e7b-8543-840f86d0f7b5\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.206457 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/373e990b-3bca-4e7b-8543-840f86d0f7b5-logs\") pod \"373e990b-3bca-4e7b-8543-840f86d0f7b5\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.207035 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/373e990b-3bca-4e7b-8543-840f86d0f7b5-logs" (OuterVolumeSpecName: "logs") pod "373e990b-3bca-4e7b-8543-840f86d0f7b5" (UID: "373e990b-3bca-4e7b-8543-840f86d0f7b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.207686 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data-custom\") pod \"373e990b-3bca-4e7b-8543-840f86d0f7b5\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.207765 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-combined-ca-bundle\") pod \"373e990b-3bca-4e7b-8543-840f86d0f7b5\" (UID: \"373e990b-3bca-4e7b-8543-840f86d0f7b5\") " Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.208272 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/373e990b-3bca-4e7b-8543-840f86d0f7b5-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.211529 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/373e990b-3bca-4e7b-8543-840f86d0f7b5-kube-api-access-hhwrd" (OuterVolumeSpecName: "kube-api-access-hhwrd") pod "373e990b-3bca-4e7b-8543-840f86d0f7b5" (UID: "373e990b-3bca-4e7b-8543-840f86d0f7b5"). InnerVolumeSpecName "kube-api-access-hhwrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.211777 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "373e990b-3bca-4e7b-8543-840f86d0f7b5" (UID: "373e990b-3bca-4e7b-8543-840f86d0f7b5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.243972 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "373e990b-3bca-4e7b-8543-840f86d0f7b5" (UID: "373e990b-3bca-4e7b-8543-840f86d0f7b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.261323 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data" (OuterVolumeSpecName: "config-data") pod "373e990b-3bca-4e7b-8543-840f86d0f7b5" (UID: "373e990b-3bca-4e7b-8543-840f86d0f7b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.278834 4775 scope.go:117] "RemoveContainer" containerID="5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d" Nov 26 06:42:27 crc kubenswrapper[4775]: E1126 06:42:27.279362 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d\": container with ID starting with 5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d not found: ID does not exist" containerID="5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.279478 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d"} err="failed to get container status \"5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d\": rpc error: code = NotFound desc = could not find container \"5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d\": container with ID starting with 5e3cd5e0acabc96a44028a71bf19c86cf4d71d5a0cb9b670c158890f10d8459d not found: ID does not exist" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.279576 4775 scope.go:117] "RemoveContainer" containerID="b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59" Nov 26 06:42:27 crc kubenswrapper[4775]: E1126 06:42:27.280178 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59\": container with ID starting with b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59 not found: ID does not exist" containerID="b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.280223 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59"} err="failed to get container status \"b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59\": rpc error: code = NotFound desc = could not find container \"b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59\": container with ID starting with b9ac3749e5990a8dfcd3b7591ff66e976a42fa3684b583e392066ff0f5e0da59 not found: ID does not exist" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.310563 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.310974 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.311004 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhwrd\" (UniqueName: \"kubernetes.io/projected/373e990b-3bca-4e7b-8543-840f86d0f7b5-kube-api-access-hhwrd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.311029 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/373e990b-3bca-4e7b-8543-840f86d0f7b5-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.480830 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5dfb875fdb-zr8j4"] Nov 26 06:42:27 crc kubenswrapper[4775]: I1126 06:42:27.492490 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5dfb875fdb-zr8j4"] Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.093153 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.148421 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerID="4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a" exitCode=0 Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.148493 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5886fff4d-zcj7s" event={"ID":"fb607340-2aa9-4b57-a4b7-a6a64457cae3","Type":"ContainerDied","Data":"4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a"} Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.151033 4775 generic.go:334] "Generic (PLEG): container finished" podID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerID="6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5" exitCode=0 Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.151088 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e769a62a-fd58-480a-ad68-010ef2490e7b","Type":"ContainerDied","Data":"6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5"} Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.151109 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e769a62a-fd58-480a-ad68-010ef2490e7b","Type":"ContainerDied","Data":"b032e0219c0addf1301050b1bf68ecfd1be50850d2dc05bcdf97ca9a8b29179b"} Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.151126 4775 scope.go:117] "RemoveContainer" containerID="c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.151234 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.184782 4775 scope.go:117] "RemoveContainer" containerID="6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.205803 4775 scope.go:117] "RemoveContainer" containerID="c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.207219 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a\": container with ID starting with c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a not found: ID does not exist" containerID="c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.207260 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a"} err="failed to get container status \"c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a\": rpc error: code = NotFound desc = could not find container \"c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a\": container with ID starting with c6b9f4f1bc6b20204ac724f4285e6bdff488491a5417ee11ca51d9e7d6d0053a not found: ID does not exist" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.207285 4775 scope.go:117] "RemoveContainer" containerID="6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.207783 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5\": container with ID starting with 6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5 not found: ID does not exist" containerID="6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.207836 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5"} err="failed to get container status \"6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5\": rpc error: code = NotFound desc = could not find container \"6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5\": container with ID starting with 6756e6a1fa3933cf5b0a76556eca2fda62a28fcf9f70bc67bee5215fe77dd1a5 not found: ID does not exist" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.226992 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65kf2\" (UniqueName: \"kubernetes.io/projected/e769a62a-fd58-480a-ad68-010ef2490e7b-kube-api-access-65kf2\") pod \"e769a62a-fd58-480a-ad68-010ef2490e7b\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.227077 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-scripts\") pod \"e769a62a-fd58-480a-ad68-010ef2490e7b\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.227148 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-combined-ca-bundle\") pod \"e769a62a-fd58-480a-ad68-010ef2490e7b\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.227257 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data\") pod \"e769a62a-fd58-480a-ad68-010ef2490e7b\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.227338 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data-custom\") pod \"e769a62a-fd58-480a-ad68-010ef2490e7b\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.227392 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e769a62a-fd58-480a-ad68-010ef2490e7b-etc-machine-id\") pod \"e769a62a-fd58-480a-ad68-010ef2490e7b\" (UID: \"e769a62a-fd58-480a-ad68-010ef2490e7b\") " Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.228040 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e769a62a-fd58-480a-ad68-010ef2490e7b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e769a62a-fd58-480a-ad68-010ef2490e7b" (UID: "e769a62a-fd58-480a-ad68-010ef2490e7b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.232107 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e769a62a-fd58-480a-ad68-010ef2490e7b" (UID: "e769a62a-fd58-480a-ad68-010ef2490e7b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.232871 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-scripts" (OuterVolumeSpecName: "scripts") pod "e769a62a-fd58-480a-ad68-010ef2490e7b" (UID: "e769a62a-fd58-480a-ad68-010ef2490e7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.242769 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e769a62a-fd58-480a-ad68-010ef2490e7b-kube-api-access-65kf2" (OuterVolumeSpecName: "kube-api-access-65kf2") pod "e769a62a-fd58-480a-ad68-010ef2490e7b" (UID: "e769a62a-fd58-480a-ad68-010ef2490e7b"). InnerVolumeSpecName "kube-api-access-65kf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.282806 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e769a62a-fd58-480a-ad68-010ef2490e7b" (UID: "e769a62a-fd58-480a-ad68-010ef2490e7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.329894 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.329928 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.329943 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.329954 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e769a62a-fd58-480a-ad68-010ef2490e7b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.329967 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65kf2\" (UniqueName: \"kubernetes.io/projected/e769a62a-fd58-480a-ad68-010ef2490e7b-kube-api-access-65kf2\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.346998 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data" (OuterVolumeSpecName: "config-data") pod "e769a62a-fd58-480a-ad68-010ef2490e7b" (UID: "e769a62a-fd58-480a-ad68-010ef2490e7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.431779 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e769a62a-fd58-480a-ad68-010ef2490e7b-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.524029 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.531835 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.549660 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550043 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon-log" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550060 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon-log" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550070 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="probe" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550076 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="probe" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550092 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="dnsmasq-dns" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550098 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="dnsmasq-dns" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550115 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550121 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550138 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="cinder-scheduler" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550144 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="cinder-scheduler" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550152 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="init" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550159 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="init" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550172 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550178 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" Nov 26 06:42:28 crc kubenswrapper[4775]: E1126 06:42:28.550192 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550197 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550349 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api-log" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550369 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon-log" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550379 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="10c96169-294e-4273-8e08-92c677dfc01c" containerName="horizon" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550391 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="cinder-scheduler" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550400 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="acbdc4be-0331-410a-8f6c-b87323d59456" containerName="dnsmasq-dns" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550408 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" containerName="probe" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.550421 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" containerName="barbican-api" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.551296 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.555164 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.563075 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.634776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.634861 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-scripts\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.634928 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8kbg\" (UniqueName: \"kubernetes.io/projected/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-kube-api-access-q8kbg\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.635036 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.635078 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-config-data\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.635102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737103 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8kbg\" (UniqueName: \"kubernetes.io/projected/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-kube-api-access-q8kbg\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737264 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-config-data\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737297 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737359 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.737549 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-scripts\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.742374 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.742506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.743535 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-config-data\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.755196 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-scripts\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.762269 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8kbg\" (UniqueName: \"kubernetes.io/projected/3bc9a6f7-ee1c-49c1-855d-362fa796cf07-kube-api-access-q8kbg\") pod \"cinder-scheduler-0\" (UID: \"3bc9a6f7-ee1c-49c1-855d-362fa796cf07\") " pod="openstack/cinder-scheduler-0" Nov 26 06:42:28 crc kubenswrapper[4775]: I1126 06:42:28.873418 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 26 06:42:29 crc kubenswrapper[4775]: W1126 06:42:29.165999 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bc9a6f7_ee1c_49c1_855d_362fa796cf07.slice/crio-5adc160db581c4bd7835d5b3ff8dc985d45f08990d14b61eb12d79ef85c53180 WatchSource:0}: Error finding container 5adc160db581c4bd7835d5b3ff8dc985d45f08990d14b61eb12d79ef85c53180: Status 404 returned error can't find the container with id 5adc160db581c4bd7835d5b3ff8dc985d45f08990d14b61eb12d79ef85c53180 Nov 26 06:42:29 crc kubenswrapper[4775]: I1126 06:42:29.187534 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 26 06:42:29 crc kubenswrapper[4775]: I1126 06:42:29.220152 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 26 06:42:29 crc kubenswrapper[4775]: I1126 06:42:29.289315 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 26 06:42:29 crc kubenswrapper[4775]: I1126 06:42:29.341325 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="373e990b-3bca-4e7b-8543-840f86d0f7b5" path="/var/lib/kubelet/pods/373e990b-3bca-4e7b-8543-840f86d0f7b5/volumes" Nov 26 06:42:29 crc kubenswrapper[4775]: I1126 06:42:29.342413 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e769a62a-fd58-480a-ad68-010ef2490e7b" path="/var/lib/kubelet/pods/e769a62a-fd58-480a-ad68-010ef2490e7b/volumes" Nov 26 06:42:29 crc kubenswrapper[4775]: I1126 06:42:29.607062 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:42:30 crc kubenswrapper[4775]: I1126 06:42:30.190478 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3bc9a6f7-ee1c-49c1-855d-362fa796cf07","Type":"ContainerStarted","Data":"b99d1f41b29e0a18ce70f33294e8f71683c3a6627fcfb41d45874439233a1200"} Nov 26 06:42:30 crc kubenswrapper[4775]: I1126 06:42:30.190520 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3bc9a6f7-ee1c-49c1-855d-362fa796cf07","Type":"ContainerStarted","Data":"5adc160db581c4bd7835d5b3ff8dc985d45f08990d14b61eb12d79ef85c53180"} Nov 26 06:42:31 crc kubenswrapper[4775]: I1126 06:42:31.153398 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7679ff5658-lwrpm" Nov 26 06:42:31 crc kubenswrapper[4775]: I1126 06:42:31.208360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3bc9a6f7-ee1c-49c1-855d-362fa796cf07","Type":"ContainerStarted","Data":"c2558fff1294b3968a3f6057b98db2fed50e465ee03ea0bd14045a2b5840444b"} Nov 26 06:42:31 crc kubenswrapper[4775]: I1126 06:42:31.258061 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.258027061 podStartE2EDuration="3.258027061s" podCreationTimestamp="2025-11-26 06:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:31.232991541 +0000 UTC m=+1034.594295493" watchObservedRunningTime="2025-11-26 06:42:31.258027061 +0000 UTC m=+1034.619331023" Nov 26 06:42:32 crc kubenswrapper[4775]: I1126 06:42:32.505857 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54d985cc65-zsv4d" Nov 26 06:42:32 crc kubenswrapper[4775]: I1126 06:42:32.637541 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5775c97bd4-547ct"] Nov 26 06:42:32 crc kubenswrapper[4775]: I1126 06:42:32.638071 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5775c97bd4-547ct" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-api" containerID="cri-o://b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91" gracePeriod=30 Nov 26 06:42:32 crc kubenswrapper[4775]: I1126 06:42:32.638348 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5775c97bd4-547ct" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-httpd" containerID="cri-o://c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0" gracePeriod=30 Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.177695 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.178987 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.181561 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ks85c" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.183259 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.184360 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.201364 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.232054 4775 generic.go:334] "Generic (PLEG): container finished" podID="116bcb10-21df-40b7-b110-10206c92ea20" containerID="c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0" exitCode=0 Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.232105 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5775c97bd4-547ct" event={"ID":"116bcb10-21df-40b7-b110-10206c92ea20","Type":"ContainerDied","Data":"c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0"} Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.337489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config-secret\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.337553 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.337613 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s95qn\" (UniqueName: \"kubernetes.io/projected/130642f3-98d0-4330-9e51-6054d1e76c98-kube-api-access-s95qn\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.337688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-combined-ca-bundle\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.440241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-combined-ca-bundle\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.440381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config-secret\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.440430 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.440480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s95qn\" (UniqueName: \"kubernetes.io/projected/130642f3-98d0-4330-9e51-6054d1e76c98-kube-api-access-s95qn\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.441615 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.447688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-combined-ca-bundle\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.471674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config-secret\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.483674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s95qn\" (UniqueName: \"kubernetes.io/projected/130642f3-98d0-4330-9e51-6054d1e76c98-kube-api-access-s95qn\") pod \"openstackclient\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.510256 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.604776 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.628936 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.640392 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.641505 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.662403 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:33 crc kubenswrapper[4775]: E1126 06:42:33.704212 4775 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 26 06:42:33 crc kubenswrapper[4775]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_130642f3-98d0-4330-9e51-6054d1e76c98_0(e9ddb26b91759e53d05e1d06be4ffa64f69dc22b77376c38e3cf19a89f9d550c): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e9ddb26b91759e53d05e1d06be4ffa64f69dc22b77376c38e3cf19a89f9d550c" Netns:"/var/run/netns/2283a599-320e-4791-9fb7-db7e98f28a4e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=e9ddb26b91759e53d05e1d06be4ffa64f69dc22b77376c38e3cf19a89f9d550c;K8S_POD_UID=130642f3-98d0-4330-9e51-6054d1e76c98" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/130642f3-98d0-4330-9e51-6054d1e76c98]: expected pod UID "130642f3-98d0-4330-9e51-6054d1e76c98" but got "3675965a-9ca7-433b-9c69-99596e74aa25" from Kube API Nov 26 06:42:33 crc kubenswrapper[4775]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 26 06:42:33 crc kubenswrapper[4775]: > Nov 26 06:42:33 crc kubenswrapper[4775]: E1126 06:42:33.704285 4775 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 26 06:42:33 crc kubenswrapper[4775]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_130642f3-98d0-4330-9e51-6054d1e76c98_0(e9ddb26b91759e53d05e1d06be4ffa64f69dc22b77376c38e3cf19a89f9d550c): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e9ddb26b91759e53d05e1d06be4ffa64f69dc22b77376c38e3cf19a89f9d550c" Netns:"/var/run/netns/2283a599-320e-4791-9fb7-db7e98f28a4e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=e9ddb26b91759e53d05e1d06be4ffa64f69dc22b77376c38e3cf19a89f9d550c;K8S_POD_UID=130642f3-98d0-4330-9e51-6054d1e76c98" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/130642f3-98d0-4330-9e51-6054d1e76c98]: expected pod UID "130642f3-98d0-4330-9e51-6054d1e76c98" but got "3675965a-9ca7-433b-9c69-99596e74aa25" from Kube API Nov 26 06:42:33 crc kubenswrapper[4775]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 26 06:42:33 crc kubenswrapper[4775]: > pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.744800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3675965a-9ca7-433b-9c69-99596e74aa25-openstack-config-secret\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.745084 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cpdp\" (UniqueName: \"kubernetes.io/projected/3675965a-9ca7-433b-9c69-99596e74aa25-kube-api-access-6cpdp\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.745231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3675965a-9ca7-433b-9c69-99596e74aa25-openstack-config\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.745295 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3675965a-9ca7-433b-9c69-99596e74aa25-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.846897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cpdp\" (UniqueName: \"kubernetes.io/projected/3675965a-9ca7-433b-9c69-99596e74aa25-kube-api-access-6cpdp\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.847009 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3675965a-9ca7-433b-9c69-99596e74aa25-openstack-config\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.847069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3675965a-9ca7-433b-9c69-99596e74aa25-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.847276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3675965a-9ca7-433b-9c69-99596e74aa25-openstack-config-secret\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.849217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3675965a-9ca7-433b-9c69-99596e74aa25-openstack-config\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.868431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3675965a-9ca7-433b-9c69-99596e74aa25-openstack-config-secret\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.868951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3675965a-9ca7-433b-9c69-99596e74aa25-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.870959 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cpdp\" (UniqueName: \"kubernetes.io/projected/3675965a-9ca7-433b-9c69-99596e74aa25-kube-api-access-6cpdp\") pod \"openstackclient\" (UID: \"3675965a-9ca7-433b-9c69-99596e74aa25\") " pod="openstack/openstackclient" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.882899 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 26 06:42:33 crc kubenswrapper[4775]: I1126 06:42:33.962145 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.240857 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.245817 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="130642f3-98d0-4330-9e51-6054d1e76c98" podUID="3675965a-9ca7-433b-9c69-99596e74aa25" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.251756 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.356581 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config\") pod \"130642f3-98d0-4330-9e51-6054d1e76c98\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.356783 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config-secret\") pod \"130642f3-98d0-4330-9e51-6054d1e76c98\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.356877 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-combined-ca-bundle\") pod \"130642f3-98d0-4330-9e51-6054d1e76c98\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.356899 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s95qn\" (UniqueName: \"kubernetes.io/projected/130642f3-98d0-4330-9e51-6054d1e76c98-kube-api-access-s95qn\") pod \"130642f3-98d0-4330-9e51-6054d1e76c98\" (UID: \"130642f3-98d0-4330-9e51-6054d1e76c98\") " Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.357126 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "130642f3-98d0-4330-9e51-6054d1e76c98" (UID: "130642f3-98d0-4330-9e51-6054d1e76c98"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.357548 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.361203 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130642f3-98d0-4330-9e51-6054d1e76c98-kube-api-access-s95qn" (OuterVolumeSpecName: "kube-api-access-s95qn") pod "130642f3-98d0-4330-9e51-6054d1e76c98" (UID: "130642f3-98d0-4330-9e51-6054d1e76c98"). InnerVolumeSpecName "kube-api-access-s95qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.361313 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "130642f3-98d0-4330-9e51-6054d1e76c98" (UID: "130642f3-98d0-4330-9e51-6054d1e76c98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.362418 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "130642f3-98d0-4330-9e51-6054d1e76c98" (UID: "130642f3-98d0-4330-9e51-6054d1e76c98"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.460493 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.460538 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/130642f3-98d0-4330-9e51-6054d1e76c98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.460549 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s95qn\" (UniqueName: \"kubernetes.io/projected/130642f3-98d0-4330-9e51-6054d1e76c98-kube-api-access-s95qn\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:34 crc kubenswrapper[4775]: I1126 06:42:34.478465 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 26 06:42:34 crc kubenswrapper[4775]: W1126 06:42:34.489073 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3675965a_9ca7_433b_9c69_99596e74aa25.slice/crio-d851c63559c645f358ce7d338afaf600f3270091e91994b472a8a52382104167 WatchSource:0}: Error finding container d851c63559c645f358ce7d338afaf600f3270091e91994b472a8a52382104167: Status 404 returned error can't find the container with id d851c63559c645f358ce7d338afaf600f3270091e91994b472a8a52382104167 Nov 26 06:42:35 crc kubenswrapper[4775]: I1126 06:42:35.248940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3675965a-9ca7-433b-9c69-99596e74aa25","Type":"ContainerStarted","Data":"d851c63559c645f358ce7d338afaf600f3270091e91994b472a8a52382104167"} Nov 26 06:42:35 crc kubenswrapper[4775]: I1126 06:42:35.248955 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 26 06:42:35 crc kubenswrapper[4775]: I1126 06:42:35.251998 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="130642f3-98d0-4330-9e51-6054d1e76c98" podUID="3675965a-9ca7-433b-9c69-99596e74aa25" Nov 26 06:42:35 crc kubenswrapper[4775]: I1126 06:42:35.343212 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130642f3-98d0-4330-9e51-6054d1e76c98" path="/var/lib/kubelet/pods/130642f3-98d0-4330-9e51-6054d1e76c98/volumes" Nov 26 06:42:36 crc kubenswrapper[4775]: I1126 06:42:36.041747 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:36 crc kubenswrapper[4775]: I1126 06:42:36.065752 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cd45b998d-gjx5d" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.639105 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6c8b9f5bf9-nsz2f"] Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.641524 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.644262 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.644429 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.644536 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.664990 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6c8b9f5bf9-nsz2f"] Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.747648 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/546c7a7f-b54b-415e-81c6-0fb3fa734e05-etc-swift\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.747694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/546c7a7f-b54b-415e-81c6-0fb3fa734e05-log-httpd\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.750842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-public-tls-certs\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.750896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/546c7a7f-b54b-415e-81c6-0fb3fa734e05-run-httpd\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.750978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-config-data\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.751113 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-combined-ca-bundle\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.751196 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khq2c\" (UniqueName: \"kubernetes.io/projected/546c7a7f-b54b-415e-81c6-0fb3fa734e05-kube-api-access-khq2c\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.751228 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-internal-tls-certs\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852440 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/546c7a7f-b54b-415e-81c6-0fb3fa734e05-etc-swift\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852474 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/546c7a7f-b54b-415e-81c6-0fb3fa734e05-log-httpd\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852496 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-public-tls-certs\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852514 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/546c7a7f-b54b-415e-81c6-0fb3fa734e05-run-httpd\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852549 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-config-data\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852603 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-combined-ca-bundle\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852638 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khq2c\" (UniqueName: \"kubernetes.io/projected/546c7a7f-b54b-415e-81c6-0fb3fa734e05-kube-api-access-khq2c\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852660 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-internal-tls-certs\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.852928 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/546c7a7f-b54b-415e-81c6-0fb3fa734e05-log-httpd\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.854201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/546c7a7f-b54b-415e-81c6-0fb3fa734e05-run-httpd\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.859887 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/546c7a7f-b54b-415e-81c6-0fb3fa734e05-etc-swift\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.860405 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.867051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-combined-ca-bundle\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.867691 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-public-tls-certs\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.868526 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-internal-tls-certs\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.870086 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/546c7a7f-b54b-415e-81c6-0fb3fa734e05-config-data\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.871840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khq2c\" (UniqueName: \"kubernetes.io/projected/546c7a7f-b54b-415e-81c6-0fb3fa734e05-kube-api-access-khq2c\") pod \"swift-proxy-6c8b9f5bf9-nsz2f\" (UID: \"546c7a7f-b54b-415e-81c6-0fb3fa734e05\") " pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:37 crc kubenswrapper[4775]: I1126 06:42:37.959202 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.057296 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-combined-ca-bundle\") pod \"116bcb10-21df-40b7-b110-10206c92ea20\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.057436 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-config\") pod \"116bcb10-21df-40b7-b110-10206c92ea20\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.057486 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-httpd-config\") pod \"116bcb10-21df-40b7-b110-10206c92ea20\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.057554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t67nr\" (UniqueName: \"kubernetes.io/projected/116bcb10-21df-40b7-b110-10206c92ea20-kube-api-access-t67nr\") pod \"116bcb10-21df-40b7-b110-10206c92ea20\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.057613 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-ovndb-tls-certs\") pod \"116bcb10-21df-40b7-b110-10206c92ea20\" (UID: \"116bcb10-21df-40b7-b110-10206c92ea20\") " Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.079295 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116bcb10-21df-40b7-b110-10206c92ea20-kube-api-access-t67nr" (OuterVolumeSpecName: "kube-api-access-t67nr") pod "116bcb10-21df-40b7-b110-10206c92ea20" (UID: "116bcb10-21df-40b7-b110-10206c92ea20"). InnerVolumeSpecName "kube-api-access-t67nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.091009 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "116bcb10-21df-40b7-b110-10206c92ea20" (UID: "116bcb10-21df-40b7-b110-10206c92ea20"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.142167 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-config" (OuterVolumeSpecName: "config") pod "116bcb10-21df-40b7-b110-10206c92ea20" (UID: "116bcb10-21df-40b7-b110-10206c92ea20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.157510 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "116bcb10-21df-40b7-b110-10206c92ea20" (UID: "116bcb10-21df-40b7-b110-10206c92ea20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.159826 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.159848 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.159901 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.159909 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t67nr\" (UniqueName: \"kubernetes.io/projected/116bcb10-21df-40b7-b110-10206c92ea20-kube-api-access-t67nr\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.176415 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "116bcb10-21df-40b7-b110-10206c92ea20" (UID: "116bcb10-21df-40b7-b110-10206c92ea20"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.262164 4775 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/116bcb10-21df-40b7-b110-10206c92ea20-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.290902 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5775c97bd4-547ct" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.290916 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5775c97bd4-547ct" event={"ID":"116bcb10-21df-40b7-b110-10206c92ea20","Type":"ContainerDied","Data":"b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91"} Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.290977 4775 scope.go:117] "RemoveContainer" containerID="c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.290973 4775 generic.go:334] "Generic (PLEG): container finished" podID="116bcb10-21df-40b7-b110-10206c92ea20" containerID="b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91" exitCode=0 Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.291027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5775c97bd4-547ct" event={"ID":"116bcb10-21df-40b7-b110-10206c92ea20","Type":"ContainerDied","Data":"5df1d6ab0c51696cf041d59af53b40fe32eb3fb139b6b321ac85e0997a283111"} Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.324006 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5775c97bd4-547ct"] Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.332025 4775 scope.go:117] "RemoveContainer" containerID="b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.332411 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5775c97bd4-547ct"] Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.352077 4775 scope.go:117] "RemoveContainer" containerID="c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0" Nov 26 06:42:38 crc kubenswrapper[4775]: E1126 06:42:38.352434 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0\": container with ID starting with c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0 not found: ID does not exist" containerID="c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.352484 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0"} err="failed to get container status \"c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0\": rpc error: code = NotFound desc = could not find container \"c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0\": container with ID starting with c1bddd21506b8b6b1f26ccc892f1093710d1dd92fde0a15d0550220949fc7ad0 not found: ID does not exist" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.352505 4775 scope.go:117] "RemoveContainer" containerID="b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91" Nov 26 06:42:38 crc kubenswrapper[4775]: E1126 06:42:38.353096 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91\": container with ID starting with b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91 not found: ID does not exist" containerID="b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.353119 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91"} err="failed to get container status \"b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91\": rpc error: code = NotFound desc = could not find container \"b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91\": container with ID starting with b979ede30b8827a8f0fca5dbe9580e7ee4f57017583b732ad024b18d88387e91 not found: ID does not exist" Nov 26 06:42:38 crc kubenswrapper[4775]: I1126 06:42:38.554374 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6c8b9f5bf9-nsz2f"] Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.074562 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.075061 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-central-agent" containerID="cri-o://c436d72e873c9a9792d542db4e5fea82b2eb902503fa297c01bc68556c2ddb37" gracePeriod=30 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.075440 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="proxy-httpd" containerID="cri-o://8e01992a6ecd9cc660a39d7675118556b27a160f4dd39c30a8a98e06bd04a641" gracePeriod=30 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.075490 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="sg-core" containerID="cri-o://ec61dec025684a387060b9d9a08a3ce4ac282766f418f44300ebb8daf241e508" gracePeriod=30 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.075549 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-notification-agent" containerID="cri-o://13559158189cc3bd1149718ea6f4241e0a4c9fa4bfc7871e682ed45f85c0db63" gracePeriod=30 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.082647 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.163:3000/\": EOF" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.107924 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.288925 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.317494 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" event={"ID":"546c7a7f-b54b-415e-81c6-0fb3fa734e05","Type":"ContainerStarted","Data":"4a57fbbff18d579162645f36edd15deb3ce2a05bd5c72374d074188e14104d12"} Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.317535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" event={"ID":"546c7a7f-b54b-415e-81c6-0fb3fa734e05","Type":"ContainerStarted","Data":"38be55d609f3d7ae3323a39f10370bfabd1183a95a51da091c00b925a0a65722"} Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.317546 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" event={"ID":"546c7a7f-b54b-415e-81c6-0fb3fa734e05","Type":"ContainerStarted","Data":"3d3dd763994170f45b5d95f717a2955b8018ab03fedfea57d41fa3b667f321ef"} Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.318432 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.318498 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.322338 4775 generic.go:334] "Generic (PLEG): container finished" podID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerID="8e01992a6ecd9cc660a39d7675118556b27a160f4dd39c30a8a98e06bd04a641" exitCode=0 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.322375 4775 generic.go:334] "Generic (PLEG): container finished" podID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerID="ec61dec025684a387060b9d9a08a3ce4ac282766f418f44300ebb8daf241e508" exitCode=2 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.322396 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerDied","Data":"8e01992a6ecd9cc660a39d7675118556b27a160f4dd39c30a8a98e06bd04a641"} Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.322419 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerDied","Data":"ec61dec025684a387060b9d9a08a3ce4ac282766f418f44300ebb8daf241e508"} Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.340107 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="116bcb10-21df-40b7-b110-10206c92ea20" path="/var/lib/kubelet/pods/116bcb10-21df-40b7-b110-10206c92ea20/volumes" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.342394 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" podStartSLOduration=2.34237367 podStartE2EDuration="2.34237367s" podCreationTimestamp="2025-11-26 06:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:39.338989895 +0000 UTC m=+1042.700293927" watchObservedRunningTime="2025-11-26 06:42:39.34237367 +0000 UTC m=+1042.703677622" Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.835313 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.835549 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-log" containerID="cri-o://a485eab588369566aa8f255e9ec6ef51f4510c4c10250579ae1223f55acd0df8" gracePeriod=30 Nov 26 06:42:39 crc kubenswrapper[4775]: I1126 06:42:39.835655 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-httpd" containerID="cri-o://778db2e53ba426fd87d3a85924c15d4815c08f416f191dd9dca0fc0dfd65600a" gracePeriod=30 Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.335597 4775 generic.go:334] "Generic (PLEG): container finished" podID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerID="a485eab588369566aa8f255e9ec6ef51f4510c4c10250579ae1223f55acd0df8" exitCode=143 Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.335685 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3883f87e-8234-4813-a3b9-774f26e1b2ea","Type":"ContainerDied","Data":"a485eab588369566aa8f255e9ec6ef51f4510c4c10250579ae1223f55acd0df8"} Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.338940 4775 generic.go:334] "Generic (PLEG): container finished" podID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerID="c436d72e873c9a9792d542db4e5fea82b2eb902503fa297c01bc68556c2ddb37" exitCode=0 Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.339764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerDied","Data":"c436d72e873c9a9792d542db4e5fea82b2eb902503fa297c01bc68556c2ddb37"} Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.666261 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.666671 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-log" containerID="cri-o://c4063b48c0ec06f1e596cb6746b388b87519e7665b6fe1c57927aead4c5db05d" gracePeriod=30 Nov 26 06:42:40 crc kubenswrapper[4775]: I1126 06:42:40.667134 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-httpd" containerID="cri-o://d93bd4dedbb36b7574fdca052093e0f1cb8b51986765480b9a021da0ea08de96" gracePeriod=30 Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.157447 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-nqxrj"] Nov 26 06:42:41 crc kubenswrapper[4775]: E1126 06:42:41.158112 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-httpd" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.158131 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-httpd" Nov 26 06:42:41 crc kubenswrapper[4775]: E1126 06:42:41.158156 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-api" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.158167 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-api" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.167915 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-api" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.167997 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="116bcb10-21df-40b7-b110-10206c92ea20" containerName="neutron-httpd" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.169168 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.186799 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nqxrj"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.243559 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-rktpq"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.245046 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.266258 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rktpq"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.276118 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8bc2-account-create-update-kqltf"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.280037 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.281739 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.289214 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8bc2-account-create-update-kqltf"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.320132 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-operator-scripts\") pod \"nova-cell0-db-create-rktpq\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.320203 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9cwn\" (UniqueName: \"kubernetes.io/projected/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-kube-api-access-l9cwn\") pod \"nova-api-db-create-nqxrj\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.320248 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-operator-scripts\") pod \"nova-api-db-create-nqxrj\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.320298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-877g4\" (UniqueName: \"kubernetes.io/projected/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-kube-api-access-877g4\") pod \"nova-cell0-db-create-rktpq\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.344530 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-v6zbv"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.345631 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.359222 4775 generic.go:334] "Generic (PLEG): container finished" podID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerID="13559158189cc3bd1149718ea6f4241e0a4c9fa4bfc7871e682ed45f85c0db63" exitCode=0 Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.359301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerDied","Data":"13559158189cc3bd1149718ea6f4241e0a4c9fa4bfc7871e682ed45f85c0db63"} Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.360382 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-v6zbv"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.364498 4775 generic.go:334] "Generic (PLEG): container finished" podID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerID="c4063b48c0ec06f1e596cb6746b388b87519e7665b6fe1c57927aead4c5db05d" exitCode=143 Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.365179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf1a7acd-833b-439f-bcb9-f83dae6dcf53","Type":"ContainerDied","Data":"c4063b48c0ec06f1e596cb6746b388b87519e7665b6fe1c57927aead4c5db05d"} Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.421689 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/331d4e96-545d-45e7-b78a-5e42bbc2bdae-operator-scripts\") pod \"nova-cell1-db-create-v6zbv\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.421782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-operator-scripts\") pod \"nova-api-db-create-nqxrj\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.421819 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/150566f8-cc6b-4ba8-b690-1962b971c9f3-operator-scripts\") pod \"nova-api-8bc2-account-create-update-kqltf\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.421882 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-877g4\" (UniqueName: \"kubernetes.io/projected/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-kube-api-access-877g4\") pod \"nova-cell0-db-create-rktpq\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.421963 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62chv\" (UniqueName: \"kubernetes.io/projected/331d4e96-545d-45e7-b78a-5e42bbc2bdae-kube-api-access-62chv\") pod \"nova-cell1-db-create-v6zbv\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.421997 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/150566f8-cc6b-4ba8-b690-1962b971c9f3-kube-api-access-f55kq\") pod \"nova-api-8bc2-account-create-update-kqltf\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.422021 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-operator-scripts\") pod \"nova-cell0-db-create-rktpq\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.422092 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9cwn\" (UniqueName: \"kubernetes.io/projected/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-kube-api-access-l9cwn\") pod \"nova-api-db-create-nqxrj\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.422583 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-operator-scripts\") pod \"nova-api-db-create-nqxrj\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.422874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-operator-scripts\") pod \"nova-cell0-db-create-rktpq\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.443453 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9cwn\" (UniqueName: \"kubernetes.io/projected/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-kube-api-access-l9cwn\") pod \"nova-api-db-create-nqxrj\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.444858 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-877g4\" (UniqueName: \"kubernetes.io/projected/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-kube-api-access-877g4\") pod \"nova-cell0-db-create-rktpq\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.459281 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-29ee-account-create-update-f556s"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.461017 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.470746 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-29ee-account-create-update-f556s"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.495988 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.496614 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.525409 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62chv\" (UniqueName: \"kubernetes.io/projected/331d4e96-545d-45e7-b78a-5e42bbc2bdae-kube-api-access-62chv\") pod \"nova-cell1-db-create-v6zbv\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.525480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/150566f8-cc6b-4ba8-b690-1962b971c9f3-kube-api-access-f55kq\") pod \"nova-api-8bc2-account-create-update-kqltf\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.525593 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/331d4e96-545d-45e7-b78a-5e42bbc2bdae-operator-scripts\") pod \"nova-cell1-db-create-v6zbv\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.525650 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/150566f8-cc6b-4ba8-b690-1962b971c9f3-operator-scripts\") pod \"nova-api-8bc2-account-create-update-kqltf\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.526501 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/150566f8-cc6b-4ba8-b690-1962b971c9f3-operator-scripts\") pod \"nova-api-8bc2-account-create-update-kqltf\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.527149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/331d4e96-545d-45e7-b78a-5e42bbc2bdae-operator-scripts\") pod \"nova-cell1-db-create-v6zbv\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.543162 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/150566f8-cc6b-4ba8-b690-1962b971c9f3-kube-api-access-f55kq\") pod \"nova-api-8bc2-account-create-update-kqltf\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.547041 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62chv\" (UniqueName: \"kubernetes.io/projected/331d4e96-545d-45e7-b78a-5e42bbc2bdae-kube-api-access-62chv\") pod \"nova-cell1-db-create-v6zbv\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.564267 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.605197 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.627818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjptz\" (UniqueName: \"kubernetes.io/projected/4cd7b78e-faa9-4232-a65c-798522553729-kube-api-access-zjptz\") pod \"nova-cell0-29ee-account-create-update-f556s\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.628539 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cd7b78e-faa9-4232-a65c-798522553729-operator-scripts\") pod \"nova-cell0-29ee-account-create-update-f556s\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.643092 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-35a8-account-create-update-tnxdk"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.644216 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.647117 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.661561 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.666614 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-35a8-account-create-update-tnxdk"] Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.731290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cd7b78e-faa9-4232-a65c-798522553729-operator-scripts\") pod \"nova-cell0-29ee-account-create-update-f556s\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.731341 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jthhv\" (UniqueName: \"kubernetes.io/projected/b546f7e2-903a-46dc-8b36-d407c66a86dc-kube-api-access-jthhv\") pod \"nova-cell1-35a8-account-create-update-tnxdk\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.731456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjptz\" (UniqueName: \"kubernetes.io/projected/4cd7b78e-faa9-4232-a65c-798522553729-kube-api-access-zjptz\") pod \"nova-cell0-29ee-account-create-update-f556s\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.731637 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b546f7e2-903a-46dc-8b36-d407c66a86dc-operator-scripts\") pod \"nova-cell1-35a8-account-create-update-tnxdk\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.737770 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cd7b78e-faa9-4232-a65c-798522553729-operator-scripts\") pod \"nova-cell0-29ee-account-create-update-f556s\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.756332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjptz\" (UniqueName: \"kubernetes.io/projected/4cd7b78e-faa9-4232-a65c-798522553729-kube-api-access-zjptz\") pod \"nova-cell0-29ee-account-create-update-f556s\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.833076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jthhv\" (UniqueName: \"kubernetes.io/projected/b546f7e2-903a-46dc-8b36-d407c66a86dc-kube-api-access-jthhv\") pod \"nova-cell1-35a8-account-create-update-tnxdk\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.833242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b546f7e2-903a-46dc-8b36-d407c66a86dc-operator-scripts\") pod \"nova-cell1-35a8-account-create-update-tnxdk\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.834089 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b546f7e2-903a-46dc-8b36-d407c66a86dc-operator-scripts\") pod \"nova-cell1-35a8-account-create-update-tnxdk\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.849141 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jthhv\" (UniqueName: \"kubernetes.io/projected/b546f7e2-903a-46dc-8b36-d407c66a86dc-kube-api-access-jthhv\") pod \"nova-cell1-35a8-account-create-update-tnxdk\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.920416 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:41 crc kubenswrapper[4775]: I1126 06:42:41.960855 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:43 crc kubenswrapper[4775]: I1126 06:42:43.381958 4775 generic.go:334] "Generic (PLEG): container finished" podID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerID="778db2e53ba426fd87d3a85924c15d4815c08f416f191dd9dca0fc0dfd65600a" exitCode=0 Nov 26 06:42:43 crc kubenswrapper[4775]: I1126 06:42:43.382160 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3883f87e-8234-4813-a3b9-774f26e1b2ea","Type":"ContainerDied","Data":"778db2e53ba426fd87d3a85924c15d4815c08f416f191dd9dca0fc0dfd65600a"} Nov 26 06:42:44 crc kubenswrapper[4775]: I1126 06:42:44.395963 4775 generic.go:334] "Generic (PLEG): container finished" podID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerID="d93bd4dedbb36b7574fdca052093e0f1cb8b51986765480b9a021da0ea08de96" exitCode=0 Nov 26 06:42:44 crc kubenswrapper[4775]: I1126 06:42:44.396009 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf1a7acd-833b-439f-bcb9-f83dae6dcf53","Type":"ContainerDied","Data":"d93bd4dedbb36b7574fdca052093e0f1cb8b51986765480b9a021da0ea08de96"} Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.255917 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.418440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3675965a-9ca7-433b-9c69-99596e74aa25","Type":"ContainerStarted","Data":"8bb1498bfc5325dd977cff2706626c6dcf655990f9a130559ebbf57607dcc544"} Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.420314 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.422180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bf1a7acd-833b-439f-bcb9-f83dae6dcf53","Type":"ContainerDied","Data":"24c95b01c91617d9af139ddc8b1200b259e1152788a6198910d2d3a225858b4a"} Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.422213 4775 scope.go:117] "RemoveContainer" containerID="d93bd4dedbb36b7574fdca052093e0f1cb8b51986765480b9a021da0ea08de96" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.422301 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427487 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-combined-ca-bundle\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427584 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7w6d\" (UniqueName: \"kubernetes.io/projected/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-kube-api-access-v7w6d\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427639 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-scripts\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427657 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-httpd-run\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427705 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-internal-tls-certs\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427758 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427779 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-logs\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.427810 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-config-data\") pod \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\" (UID: \"bf1a7acd-833b-439f-bcb9-f83dae6dcf53\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.431333 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.436917 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-kube-api-access-v7w6d" (OuterVolumeSpecName: "kube-api-access-v7w6d") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "kube-api-access-v7w6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.438914 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-logs" (OuterVolumeSpecName: "logs") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.440352 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4","Type":"ContainerDied","Data":"8b00f2498c4b78cf4c91aca8036a14b0752d66bce79c755506852c7bfe5ec650"} Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.440471 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.440496 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-scripts" (OuterVolumeSpecName: "scripts") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.457556 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.055053311 podStartE2EDuration="12.457533395s" podCreationTimestamp="2025-11-26 06:42:33 +0000 UTC" firstStartedPulling="2025-11-26 06:42:34.494425705 +0000 UTC m=+1037.855729687" lastFinishedPulling="2025-11-26 06:42:44.896905819 +0000 UTC m=+1048.258209771" observedRunningTime="2025-11-26 06:42:45.430891456 +0000 UTC m=+1048.792195408" watchObservedRunningTime="2025-11-26 06:42:45.457533395 +0000 UTC m=+1048.818837347" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.466107 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.487347 4775 scope.go:117] "RemoveContainer" containerID="c4063b48c0ec06f1e596cb6746b388b87519e7665b6fe1c57927aead4c5db05d" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.509175 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-v6zbv"] Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.519950 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-35a8-account-create-update-tnxdk"] Nov 26 06:42:45 crc kubenswrapper[4775]: W1126 06:42:45.522971 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod331d4e96_545d_45e7_b78a_5e42bbc2bdae.slice/crio-468bf2a8361a4477c5d6da2caa43186c800e4f05c9263a33a18a70e592f9a979 WatchSource:0}: Error finding container 468bf2a8361a4477c5d6da2caa43186c800e4f05c9263a33a18a70e592f9a979: Status 404 returned error can't find the container with id 468bf2a8361a4477c5d6da2caa43186c800e4f05c9263a33a18a70e592f9a979 Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.530143 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-combined-ca-bundle\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.530182 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-scripts\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531064 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-run-httpd\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpqs5\" (UniqueName: \"kubernetes.io/projected/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-kube-api-access-jpqs5\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531167 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-sg-core-conf-yaml\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-log-httpd\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531269 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-config-data\") pod \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\" (UID: \"26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531763 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531780 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531790 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7w6d\" (UniqueName: \"kubernetes.io/projected/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-kube-api-access-v7w6d\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531799 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.531807 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.533294 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.533404 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.533456 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.534197 4775 scope.go:117] "RemoveContainer" containerID="8e01992a6ecd9cc660a39d7675118556b27a160f4dd39c30a8a98e06bd04a641" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.537169 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-scripts" (OuterVolumeSpecName: "scripts") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.543195 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-kube-api-access-jpqs5" (OuterVolumeSpecName: "kube-api-access-jpqs5") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "kube-api-access-jpqs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.595465 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.622009 4775 scope.go:117] "RemoveContainer" containerID="ec61dec025684a387060b9d9a08a3ce4ac282766f418f44300ebb8daf241e508" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.631874 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-config-data" (OuterVolumeSpecName: "config-data") pod "bf1a7acd-833b-439f-bcb9-f83dae6dcf53" (UID: "bf1a7acd-833b-439f-bcb9-f83dae6dcf53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633043 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633058 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633068 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633077 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpqs5\" (UniqueName: \"kubernetes.io/projected/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-kube-api-access-jpqs5\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633087 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633095 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.633103 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1a7acd-833b-439f-bcb9-f83dae6dcf53-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.634998 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.647814 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.679552 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.702696 4775 scope.go:117] "RemoveContainer" containerID="13559158189cc3bd1149718ea6f4241e0a4c9fa4bfc7871e682ed45f85c0db63" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.736386 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-combined-ca-bundle\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.736440 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-config-data\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.736479 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-public-tls-certs\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.736499 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27g4s\" (UniqueName: \"kubernetes.io/projected/3883f87e-8234-4813-a3b9-774f26e1b2ea-kube-api-access-27g4s\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.737176 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.737221 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-scripts\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.737271 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-logs\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.737294 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-httpd-run\") pod \"3883f87e-8234-4813-a3b9-774f26e1b2ea\" (UID: \"3883f87e-8234-4813-a3b9-774f26e1b2ea\") " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.737676 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.737688 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.738906 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.770033 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-logs" (OuterVolumeSpecName: "logs") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.791613 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3883f87e-8234-4813-a3b9-774f26e1b2ea-kube-api-access-27g4s" (OuterVolumeSpecName: "kube-api-access-27g4s") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "kube-api-access-27g4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.791813 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.803768 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-scripts" (OuterVolumeSpecName: "scripts") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.839861 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27g4s\" (UniqueName: \"kubernetes.io/projected/3883f87e-8234-4813-a3b9-774f26e1b2ea-kube-api-access-27g4s\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.840163 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.840175 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.840184 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.840194 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3883f87e-8234-4813-a3b9-774f26e1b2ea-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.923981 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-config-data" (OuterVolumeSpecName: "config-data") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.927884 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8bc2-account-create-update-kqltf"] Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.948568 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.949462 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.972287 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.973942 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.988405 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nqxrj"] Nov 26 06:42:45 crc kubenswrapper[4775]: I1126 06:42:45.988568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3883f87e-8234-4813-a3b9-774f26e1b2ea" (UID: "3883f87e-8234-4813-a3b9-774f26e1b2ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.001176 4775 scope.go:117] "RemoveContainer" containerID="c436d72e873c9a9792d542db4e5fea82b2eb902503fa297c01bc68556c2ddb37" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.005147 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-config-data" (OuterVolumeSpecName: "config-data") pod "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" (UID: "26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.005291 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rktpq"] Nov 26 06:42:46 crc kubenswrapper[4775]: W1126 06:42:46.020222 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cd7b78e_faa9_4232_a65c_798522553729.slice/crio-503ab583349dbaf9b6e27b693a3f8c6c3f1375c54d0c54b7bab508c3895cb0f4 WatchSource:0}: Error finding container 503ab583349dbaf9b6e27b693a3f8c6c3f1375c54d0c54b7bab508c3895cb0f4: Status 404 returned error can't find the container with id 503ab583349dbaf9b6e27b693a3f8c6c3f1375c54d0c54b7bab508c3895cb0f4 Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.024969 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-29ee-account-create-update-f556s"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.057475 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.057529 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.057543 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.057558 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.057571 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3883f87e-8234-4813-a3b9-774f26e1b2ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.062812 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.070621 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.078701 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079109 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079130 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079143 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079151 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079171 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-central-agent" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079177 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-central-agent" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079191 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-notification-agent" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079198 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-notification-agent" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079208 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="proxy-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079215 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="proxy-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079236 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-log" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079243 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-log" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079269 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-log" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079276 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-log" Nov 26 06:42:46 crc kubenswrapper[4775]: E1126 06:42:46.079291 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="sg-core" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079299 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="sg-core" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079501 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-log" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079516 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="proxy-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079528 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="sg-core" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079542 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" containerName="glance-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079559 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-central-agent" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079571 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" containerName="ceilometer-notification-agent" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079580 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-log" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.079591 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" containerName="glance-httpd" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.081081 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.083749 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.100330 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.100540 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.150697 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159573 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159610 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159634 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46819d15-5e66-4501-b2f0-570f3b768207-logs\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159726 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cdjw\" (UniqueName: \"kubernetes.io/projected/46819d15-5e66-4501-b2f0-570f3b768207-kube-api-access-9cdjw\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.159776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46819d15-5e66-4501-b2f0-570f3b768207-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.166262 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.176507 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.180322 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.182458 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.182758 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.195105 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266588 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z46rx\" (UniqueName: \"kubernetes.io/projected/6ca8e03f-851c-4196-aeb7-f153d60ee295-kube-api-access-z46rx\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266655 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-log-httpd\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-scripts\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-config-data\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266820 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266845 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-run-httpd\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46819d15-5e66-4501-b2f0-570f3b768207-logs\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.266973 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cdjw\" (UniqueName: \"kubernetes.io/projected/46819d15-5e66-4501-b2f0-570f3b768207-kube-api-access-9cdjw\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267116 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46819d15-5e66-4501-b2f0-570f3b768207-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267624 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/46819d15-5e66-4501-b2f0-570f3b768207-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267833 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.267848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46819d15-5e66-4501-b2f0-570f3b768207-logs\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.272186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-scripts\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.272699 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.272890 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-config-data\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.273355 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46819d15-5e66-4501-b2f0-570f3b768207-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.294462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cdjw\" (UniqueName: \"kubernetes.io/projected/46819d15-5e66-4501-b2f0-570f3b768207-kube-api-access-9cdjw\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.313841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"46819d15-5e66-4501-b2f0-570f3b768207\") " pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368079 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-scripts\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368122 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368161 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-config-data\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368185 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368205 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-run-httpd\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368310 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z46rx\" (UniqueName: \"kubernetes.io/projected/6ca8e03f-851c-4196-aeb7-f153d60ee295-kube-api-access-z46rx\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-log-httpd\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.368773 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-log-httpd\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.369030 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-run-httpd\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.371986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.377572 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-config-data\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.379352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-scripts\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.386277 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.409264 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z46rx\" (UniqueName: \"kubernetes.io/projected/6ca8e03f-851c-4196-aeb7-f153d60ee295-kube-api-access-z46rx\") pod \"ceilometer-0\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.430116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.460345 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-v6zbv" event={"ID":"331d4e96-545d-45e7-b78a-5e42bbc2bdae","Type":"ContainerStarted","Data":"1494c44d9e6c127d6fd7a4c3bb522cfeb9b3129474bbcdbe74e837e7b119498b"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.460398 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-v6zbv" event={"ID":"331d4e96-545d-45e7-b78a-5e42bbc2bdae","Type":"ContainerStarted","Data":"468bf2a8361a4477c5d6da2caa43186c800e4f05c9263a33a18a70e592f9a979"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.465628 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" event={"ID":"b546f7e2-903a-46dc-8b36-d407c66a86dc","Type":"ContainerStarted","Data":"8573f5ed0b5ab91cf747fc16180f4048b561736f01118aa510a0e58d8db359db"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.465823 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" event={"ID":"b546f7e2-903a-46dc-8b36-d407c66a86dc","Type":"ContainerStarted","Data":"63a013da750b05cd5714976152889b1da2513c232cbce81f6cef52600f516cf4"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.485394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nqxrj" event={"ID":"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8","Type":"ContainerStarted","Data":"309a0197f97657300cf97c1cbd375dbd0b23b1f8bf011ece5daa15a89ca7a48a"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.485448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nqxrj" event={"ID":"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8","Type":"ContainerStarted","Data":"4cf152e88e4e3a764dd1cb4336a1a474298ea406f783f0611bb3a197c0e38c4d"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.490624 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-v6zbv" podStartSLOduration=5.49060977 podStartE2EDuration="5.49060977s" podCreationTimestamp="2025-11-26 06:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:46.482190129 +0000 UTC m=+1049.843494081" watchObservedRunningTime="2025-11-26 06:42:46.49060977 +0000 UTC m=+1049.851913732" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.505041 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rktpq" event={"ID":"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b","Type":"ContainerStarted","Data":"be932c6ad24b9c6d450bd9d0ed52ecafef3127d7bc1d1cc703f0b7f3a6c5ca9b"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.505086 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rktpq" event={"ID":"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b","Type":"ContainerStarted","Data":"ef57c4702e0ad973dd576b20c166684062d7dbd0f144f12c2d8089db126e11c7"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.506698 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" podStartSLOduration=5.506686954 podStartE2EDuration="5.506686954s" podCreationTimestamp="2025-11-26 06:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:46.506639563 +0000 UTC m=+1049.867943515" watchObservedRunningTime="2025-11-26 06:42:46.506686954 +0000 UTC m=+1049.867990906" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.532154 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.534221 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-nqxrj" podStartSLOduration=5.534200205 podStartE2EDuration="5.534200205s" podCreationTimestamp="2025-11-26 06:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:46.532977704 +0000 UTC m=+1049.894281676" watchObservedRunningTime="2025-11-26 06:42:46.534200205 +0000 UTC m=+1049.895504147" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.539936 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3883f87e-8234-4813-a3b9-774f26e1b2ea","Type":"ContainerDied","Data":"27332bf08d0f217b4aa0312e4ce04d0256c4721a3e03bbe63d605d4a7713144b"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.540089 4775 scope.go:117] "RemoveContainer" containerID="778db2e53ba426fd87d3a85924c15d4815c08f416f191dd9dca0fc0dfd65600a" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.540329 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.552956 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-29ee-account-create-update-f556s" event={"ID":"4cd7b78e-faa9-4232-a65c-798522553729","Type":"ContainerStarted","Data":"887a3ba4b667f45b7ccd86cf9e4a1e98844e5fc6b934cdb72c21b630df9cae36"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.553016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-29ee-account-create-update-f556s" event={"ID":"4cd7b78e-faa9-4232-a65c-798522553729","Type":"ContainerStarted","Data":"503ab583349dbaf9b6e27b693a3f8c6c3f1375c54d0c54b7bab508c3895cb0f4"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.560440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8bc2-account-create-update-kqltf" event={"ID":"150566f8-cc6b-4ba8-b690-1962b971c9f3","Type":"ContainerStarted","Data":"1d07ddf20f2967d4dd924b1542743dfc7014c9e95ea705ec23cd638c520995d2"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.560576 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8bc2-account-create-update-kqltf" event={"ID":"150566f8-cc6b-4ba8-b690-1962b971c9f3","Type":"ContainerStarted","Data":"5625a1478abc5849332356d624ed633b2ae51403b9ae2dcf31cedc6e1a622dd7"} Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.585881 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-rktpq" podStartSLOduration=5.585865712 podStartE2EDuration="5.585865712s" podCreationTimestamp="2025-11-26 06:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:46.559616333 +0000 UTC m=+1049.920920305" watchObservedRunningTime="2025-11-26 06:42:46.585865712 +0000 UTC m=+1049.947169664" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.610470 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-29ee-account-create-update-f556s" podStartSLOduration=5.61045453 podStartE2EDuration="5.61045453s" podCreationTimestamp="2025-11-26 06:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:46.593939655 +0000 UTC m=+1049.955243607" watchObservedRunningTime="2025-11-26 06:42:46.61045453 +0000 UTC m=+1049.971758482" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.621916 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-8bc2-account-create-update-kqltf" podStartSLOduration=5.621899147 podStartE2EDuration="5.621899147s" podCreationTimestamp="2025-11-26 06:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:46.609361862 +0000 UTC m=+1049.970665834" watchObservedRunningTime="2025-11-26 06:42:46.621899147 +0000 UTC m=+1049.983203099" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.655923 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.660061 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.666961 4775 scope.go:117] "RemoveContainer" containerID="a485eab588369566aa8f255e9ec6ef51f4510c4c10250579ae1223f55acd0df8" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.678758 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.680602 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.690625 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.690923 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.691245 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.783549 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.783830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-scripts\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.783874 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmclp\" (UniqueName: \"kubernetes.io/projected/58818a45-1408-4416-b18e-814acd04b059-kube-api-access-jmclp\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.783906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58818a45-1408-4416-b18e-814acd04b059-logs\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.783931 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-config-data\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.784009 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.784027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.784139 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58818a45-1408-4416-b18e-814acd04b059-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888401 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmclp\" (UniqueName: \"kubernetes.io/projected/58818a45-1408-4416-b18e-814acd04b059-kube-api-access-jmclp\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888445 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58818a45-1408-4416-b18e-814acd04b059-logs\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-config-data\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58818a45-1408-4416-b18e-814acd04b059-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888647 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.888668 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-scripts\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.890145 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.891145 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58818a45-1408-4416-b18e-814acd04b059-logs\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.893660 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58818a45-1408-4416-b18e-814acd04b059-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.901232 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-scripts\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.909889 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-config-data\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.913237 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmclp\" (UniqueName: \"kubernetes.io/projected/58818a45-1408-4416-b18e-814acd04b059-kube-api-access-jmclp\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.914411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.920342 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58818a45-1408-4416-b18e-814acd04b059-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:46 crc kubenswrapper[4775]: I1126 06:42:46.971647 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"58818a45-1408-4416-b18e-814acd04b059\") " pod="openstack/glance-default-external-api-0" Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.040357 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.154664 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.184751 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:47 crc kubenswrapper[4775]: W1126 06:42:47.236392 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ca8e03f_851c_4196_aeb7_f153d60ee295.slice/crio-08602dd948b1ddae5654bcab13bef13a0cc86365102cd5a7766ef40490ec6064 WatchSource:0}: Error finding container 08602dd948b1ddae5654bcab13bef13a0cc86365102cd5a7766ef40490ec6064: Status 404 returned error can't find the container with id 08602dd948b1ddae5654bcab13bef13a0cc86365102cd5a7766ef40490ec6064 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.343493 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4" path="/var/lib/kubelet/pods/26e75b6d-2cb2-44d0-9633-3c4fb5c17dd4/volumes" Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.344967 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3883f87e-8234-4813-a3b9-774f26e1b2ea" path="/var/lib/kubelet/pods/3883f87e-8234-4813-a3b9-774f26e1b2ea/volumes" Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.346017 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf1a7acd-833b-439f-bcb9-f83dae6dcf53" path="/var/lib/kubelet/pods/bf1a7acd-833b-439f-bcb9-f83dae6dcf53/volumes" Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.589857 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46819d15-5e66-4501-b2f0-570f3b768207","Type":"ContainerStarted","Data":"cf1edfd88515a8b7470ef4d0a1e5258f5da93d080c4f467c1d97ac0a483876e2"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.594925 4775 generic.go:334] "Generic (PLEG): container finished" podID="150566f8-cc6b-4ba8-b690-1962b971c9f3" containerID="1d07ddf20f2967d4dd924b1542743dfc7014c9e95ea705ec23cd638c520995d2" exitCode=0 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.595297 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8bc2-account-create-update-kqltf" event={"ID":"150566f8-cc6b-4ba8-b690-1962b971c9f3","Type":"ContainerDied","Data":"1d07ddf20f2967d4dd924b1542743dfc7014c9e95ea705ec23cd638c520995d2"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.599893 4775 generic.go:334] "Generic (PLEG): container finished" podID="b546f7e2-903a-46dc-8b36-d407c66a86dc" containerID="8573f5ed0b5ab91cf747fc16180f4048b561736f01118aa510a0e58d8db359db" exitCode=0 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.599945 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" event={"ID":"b546f7e2-903a-46dc-8b36-d407c66a86dc","Type":"ContainerDied","Data":"8573f5ed0b5ab91cf747fc16180f4048b561736f01118aa510a0e58d8db359db"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.602981 4775 generic.go:334] "Generic (PLEG): container finished" podID="5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" containerID="be932c6ad24b9c6d450bd9d0ed52ecafef3127d7bc1d1cc703f0b7f3a6c5ca9b" exitCode=0 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.603090 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rktpq" event={"ID":"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b","Type":"ContainerDied","Data":"be932c6ad24b9c6d450bd9d0ed52ecafef3127d7bc1d1cc703f0b7f3a6c5ca9b"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.604793 4775 generic.go:334] "Generic (PLEG): container finished" podID="d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" containerID="309a0197f97657300cf97c1cbd375dbd0b23b1f8bf011ece5daa15a89ca7a48a" exitCode=0 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.604910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nqxrj" event={"ID":"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8","Type":"ContainerDied","Data":"309a0197f97657300cf97c1cbd375dbd0b23b1f8bf011ece5daa15a89ca7a48a"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.611377 4775 generic.go:334] "Generic (PLEG): container finished" podID="331d4e96-545d-45e7-b78a-5e42bbc2bdae" containerID="1494c44d9e6c127d6fd7a4c3bb522cfeb9b3129474bbcdbe74e837e7b119498b" exitCode=0 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.611446 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-v6zbv" event={"ID":"331d4e96-545d-45e7-b78a-5e42bbc2bdae","Type":"ContainerDied","Data":"1494c44d9e6c127d6fd7a4c3bb522cfeb9b3129474bbcdbe74e837e7b119498b"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.613666 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerStarted","Data":"08602dd948b1ddae5654bcab13bef13a0cc86365102cd5a7766ef40490ec6064"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.616515 4775 generic.go:334] "Generic (PLEG): container finished" podID="4cd7b78e-faa9-4232-a65c-798522553729" containerID="887a3ba4b667f45b7ccd86cf9e4a1e98844e5fc6b934cdb72c21b630df9cae36" exitCode=0 Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.616545 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-29ee-account-create-update-f556s" event={"ID":"4cd7b78e-faa9-4232-a65c-798522553729","Type":"ContainerDied","Data":"887a3ba4b667f45b7ccd86cf9e4a1e98844e5fc6b934cdb72c21b630df9cae36"} Nov 26 06:42:47 crc kubenswrapper[4775]: I1126 06:42:47.708202 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.013836 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.110351 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6c8b9f5bf9-nsz2f" Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.631958 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerStarted","Data":"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504"} Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.636174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"58818a45-1408-4416-b18e-814acd04b059","Type":"ContainerStarted","Data":"045a3f868026a23a4354e01efefa6b50e0463015b6692b2837c505cf299ef135"} Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.636258 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"58818a45-1408-4416-b18e-814acd04b059","Type":"ContainerStarted","Data":"d1ccb4b7543729404d297c74b1413776743b373d33275ae37e9d02accd3991cf"} Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.639495 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46819d15-5e66-4501-b2f0-570f3b768207","Type":"ContainerStarted","Data":"1c02d9f1524c29a86416c3bc91fbc24eb46769840a17c2fa983c23b2203e5ec5"} Nov 26 06:42:48 crc kubenswrapper[4775]: I1126 06:42:48.968058 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.079220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b546f7e2-903a-46dc-8b36-d407c66a86dc-operator-scripts\") pod \"b546f7e2-903a-46dc-8b36-d407c66a86dc\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.079455 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jthhv\" (UniqueName: \"kubernetes.io/projected/b546f7e2-903a-46dc-8b36-d407c66a86dc-kube-api-access-jthhv\") pod \"b546f7e2-903a-46dc-8b36-d407c66a86dc\" (UID: \"b546f7e2-903a-46dc-8b36-d407c66a86dc\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.079851 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b546f7e2-903a-46dc-8b36-d407c66a86dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b546f7e2-903a-46dc-8b36-d407c66a86dc" (UID: "b546f7e2-903a-46dc-8b36-d407c66a86dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.082341 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b546f7e2-903a-46dc-8b36-d407c66a86dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.083881 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.087306 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b546f7e2-903a-46dc-8b36-d407c66a86dc-kube-api-access-jthhv" (OuterVolumeSpecName: "kube-api-access-jthhv") pod "b546f7e2-903a-46dc-8b36-d407c66a86dc" (UID: "b546f7e2-903a-46dc-8b36-d407c66a86dc"). InnerVolumeSpecName "kube-api-access-jthhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.183947 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62chv\" (UniqueName: \"kubernetes.io/projected/331d4e96-545d-45e7-b78a-5e42bbc2bdae-kube-api-access-62chv\") pod \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.184167 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/331d4e96-545d-45e7-b78a-5e42bbc2bdae-operator-scripts\") pod \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\" (UID: \"331d4e96-545d-45e7-b78a-5e42bbc2bdae\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.184673 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jthhv\" (UniqueName: \"kubernetes.io/projected/b546f7e2-903a-46dc-8b36-d407c66a86dc-kube-api-access-jthhv\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.186118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/331d4e96-545d-45e7-b78a-5e42bbc2bdae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "331d4e96-545d-45e7-b78a-5e42bbc2bdae" (UID: "331d4e96-545d-45e7-b78a-5e42bbc2bdae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.190957 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331d4e96-545d-45e7-b78a-5e42bbc2bdae-kube-api-access-62chv" (OuterVolumeSpecName: "kube-api-access-62chv") pod "331d4e96-545d-45e7-b78a-5e42bbc2bdae" (UID: "331d4e96-545d-45e7-b78a-5e42bbc2bdae"). InnerVolumeSpecName "kube-api-access-62chv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.239161 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.283615 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.288082 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/331d4e96-545d-45e7-b78a-5e42bbc2bdae-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.288124 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62chv\" (UniqueName: \"kubernetes.io/projected/331d4e96-545d-45e7-b78a-5e42bbc2bdae-kube-api-access-62chv\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.288283 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5886fff4d-zcj7s" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.288366 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.288630 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.313572 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.389788 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-operator-scripts\") pod \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.389859 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cd7b78e-faa9-4232-a65c-798522553729-operator-scripts\") pod \"4cd7b78e-faa9-4232-a65c-798522553729\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.389896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-operator-scripts\") pod \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.389916 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/150566f8-cc6b-4ba8-b690-1962b971c9f3-kube-api-access-f55kq\") pod \"150566f8-cc6b-4ba8-b690-1962b971c9f3\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.389944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-877g4\" (UniqueName: \"kubernetes.io/projected/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-kube-api-access-877g4\") pod \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\" (UID: \"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.389977 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/150566f8-cc6b-4ba8-b690-1962b971c9f3-operator-scripts\") pod \"150566f8-cc6b-4ba8-b690-1962b971c9f3\" (UID: \"150566f8-cc6b-4ba8-b690-1962b971c9f3\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.390048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjptz\" (UniqueName: \"kubernetes.io/projected/4cd7b78e-faa9-4232-a65c-798522553729-kube-api-access-zjptz\") pod \"4cd7b78e-faa9-4232-a65c-798522553729\" (UID: \"4cd7b78e-faa9-4232-a65c-798522553729\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.390116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9cwn\" (UniqueName: \"kubernetes.io/projected/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-kube-api-access-l9cwn\") pod \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\" (UID: \"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8\") " Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.391180 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" (UID: "5a2950b4-e9c2-45f3-9cdc-4bace1b0388b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.391591 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd7b78e-faa9-4232-a65c-798522553729-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4cd7b78e-faa9-4232-a65c-798522553729" (UID: "4cd7b78e-faa9-4232-a65c-798522553729"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.391911 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" (UID: "d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.394823 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/150566f8-cc6b-4ba8-b690-1962b971c9f3-kube-api-access-f55kq" (OuterVolumeSpecName: "kube-api-access-f55kq") pod "150566f8-cc6b-4ba8-b690-1962b971c9f3" (UID: "150566f8-cc6b-4ba8-b690-1962b971c9f3"). InnerVolumeSpecName "kube-api-access-f55kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.395171 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/150566f8-cc6b-4ba8-b690-1962b971c9f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "150566f8-cc6b-4ba8-b690-1962b971c9f3" (UID: "150566f8-cc6b-4ba8-b690-1962b971c9f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.401694 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-kube-api-access-877g4" (OuterVolumeSpecName: "kube-api-access-877g4") pod "5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" (UID: "5a2950b4-e9c2-45f3-9cdc-4bace1b0388b"). InnerVolumeSpecName "kube-api-access-877g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.403355 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd7b78e-faa9-4232-a65c-798522553729-kube-api-access-zjptz" (OuterVolumeSpecName: "kube-api-access-zjptz") pod "4cd7b78e-faa9-4232-a65c-798522553729" (UID: "4cd7b78e-faa9-4232-a65c-798522553729"). InnerVolumeSpecName "kube-api-access-zjptz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.410175 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-kube-api-access-l9cwn" (OuterVolumeSpecName: "kube-api-access-l9cwn") pod "d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" (UID: "d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8"). InnerVolumeSpecName "kube-api-access-l9cwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.412673 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492336 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/150566f8-cc6b-4ba8-b690-1962b971c9f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492372 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjptz\" (UniqueName: \"kubernetes.io/projected/4cd7b78e-faa9-4232-a65c-798522553729-kube-api-access-zjptz\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492385 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9cwn\" (UniqueName: \"kubernetes.io/projected/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-kube-api-access-l9cwn\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492394 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492402 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cd7b78e-faa9-4232-a65c-798522553729-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492410 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492418 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/150566f8-cc6b-4ba8-b690-1962b971c9f3-kube-api-access-f55kq\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.492425 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-877g4\" (UniqueName: \"kubernetes.io/projected/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b-kube-api-access-877g4\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.653384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"46819d15-5e66-4501-b2f0-570f3b768207","Type":"ContainerStarted","Data":"a003a740e3445c30c47ca98edf7b3bd0fc92f34b501cc47619c0bf3eb48b56bb"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.656425 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-29ee-account-create-update-f556s" event={"ID":"4cd7b78e-faa9-4232-a65c-798522553729","Type":"ContainerDied","Data":"503ab583349dbaf9b6e27b693a3f8c6c3f1375c54d0c54b7bab508c3895cb0f4"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.656460 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="503ab583349dbaf9b6e27b693a3f8c6c3f1375c54d0c54b7bab508c3895cb0f4" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.656514 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-29ee-account-create-update-f556s" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.672659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" event={"ID":"b546f7e2-903a-46dc-8b36-d407c66a86dc","Type":"ContainerDied","Data":"63a013da750b05cd5714976152889b1da2513c232cbce81f6cef52600f516cf4"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.672708 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a013da750b05cd5714976152889b1da2513c232cbce81f6cef52600f516cf4" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.672792 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-35a8-account-create-update-tnxdk" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.680622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerStarted","Data":"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.684485 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rktpq" event={"ID":"5a2950b4-e9c2-45f3-9cdc-4bace1b0388b","Type":"ContainerDied","Data":"ef57c4702e0ad973dd576b20c166684062d7dbd0f144f12c2d8089db126e11c7"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.684534 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef57c4702e0ad973dd576b20c166684062d7dbd0f144f12c2d8089db126e11c7" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.684587 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.684570891 podStartE2EDuration="3.684570891s" podCreationTimestamp="2025-11-26 06:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:49.679765161 +0000 UTC m=+1053.041069133" watchObservedRunningTime="2025-11-26 06:42:49.684570891 +0000 UTC m=+1053.045874863" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.684622 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rktpq" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.687382 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nqxrj" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.687430 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nqxrj" event={"ID":"d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8","Type":"ContainerDied","Data":"4cf152e88e4e3a764dd1cb4336a1a474298ea406f783f0611bb3a197c0e38c4d"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.687467 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cf152e88e4e3a764dd1cb4336a1a474298ea406f783f0611bb3a197c0e38c4d" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.695279 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8bc2-account-create-update-kqltf" event={"ID":"150566f8-cc6b-4ba8-b690-1962b971c9f3","Type":"ContainerDied","Data":"5625a1478abc5849332356d624ed633b2ae51403b9ae2dcf31cedc6e1a622dd7"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.695319 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5625a1478abc5849332356d624ed633b2ae51403b9ae2dcf31cedc6e1a622dd7" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.695393 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8bc2-account-create-update-kqltf" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.700373 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-v6zbv" event={"ID":"331d4e96-545d-45e7-b78a-5e42bbc2bdae","Type":"ContainerDied","Data":"468bf2a8361a4477c5d6da2caa43186c800e4f05c9263a33a18a70e592f9a979"} Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.700414 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468bf2a8361a4477c5d6da2caa43186c800e4f05c9263a33a18a70e592f9a979" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.700466 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-v6zbv" Nov 26 06:42:49 crc kubenswrapper[4775]: I1126 06:42:49.715107 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"58818a45-1408-4416-b18e-814acd04b059","Type":"ContainerStarted","Data":"b3a6a6e17f5149c14b46074d12b6b7bda86da0bfa0596ca28a99f1e0beb0fb31"} Nov 26 06:42:50 crc kubenswrapper[4775]: I1126 06:42:50.727155 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerStarted","Data":"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5"} Nov 26 06:42:50 crc kubenswrapper[4775]: I1126 06:42:50.747326 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.747307414 podStartE2EDuration="4.747307414s" podCreationTimestamp="2025-11-26 06:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:42:50.746077443 +0000 UTC m=+1054.107381405" watchObservedRunningTime="2025-11-26 06:42:50.747307414 +0000 UTC m=+1054.108611376" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.420435 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.420512 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756199 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bdd6t"] Nov 26 06:42:51 crc kubenswrapper[4775]: E1126 06:42:51.756543 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cd7b78e-faa9-4232-a65c-798522553729" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756554 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cd7b78e-faa9-4232-a65c-798522553729" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: E1126 06:42:51.756568 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b546f7e2-903a-46dc-8b36-d407c66a86dc" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756574 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b546f7e2-903a-46dc-8b36-d407c66a86dc" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: E1126 06:42:51.756585 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756591 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: E1126 06:42:51.756605 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331d4e96-545d-45e7-b78a-5e42bbc2bdae" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756610 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="331d4e96-545d-45e7-b78a-5e42bbc2bdae" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: E1126 06:42:51.756627 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756633 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: E1126 06:42:51.756652 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150566f8-cc6b-4ba8-b690-1962b971c9f3" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756659 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="150566f8-cc6b-4ba8-b690-1962b971c9f3" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756841 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="150566f8-cc6b-4ba8-b690-1962b971c9f3" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756854 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="331d4e96-545d-45e7-b78a-5e42bbc2bdae" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756860 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b546f7e2-903a-46dc-8b36-d407c66a86dc" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756874 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756882 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" containerName="mariadb-database-create" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.756899 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cd7b78e-faa9-4232-a65c-798522553729" containerName="mariadb-account-create-update" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.757422 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.759548 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.759699 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.761249 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6bjxj" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.791645 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bdd6t"] Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.932448 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-scripts\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.932698 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-config-data\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.932741 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzj8\" (UniqueName: \"kubernetes.io/projected/1982e34b-5fe7-4993-88dc-418d7786d001-kube-api-access-2nzj8\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:51 crc kubenswrapper[4775]: I1126 06:42:51.932762 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.034586 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-scripts\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.034649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-config-data\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.034676 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzj8\" (UniqueName: \"kubernetes.io/projected/1982e34b-5fe7-4993-88dc-418d7786d001-kube-api-access-2nzj8\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.034696 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.040657 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.042506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-config-data\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.049177 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-scripts\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.059918 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzj8\" (UniqueName: \"kubernetes.io/projected/1982e34b-5fe7-4993-88dc-418d7786d001-kube-api-access-2nzj8\") pod \"nova-cell0-conductor-db-sync-bdd6t\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.082761 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:42:52 crc kubenswrapper[4775]: W1126 06:42:52.586131 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1982e34b_5fe7_4993_88dc_418d7786d001.slice/crio-73f38fa7ca820bc3b6e0c00b47da84dfc282ca0ce1e4749c7253d7d3d3ae06b7 WatchSource:0}: Error finding container 73f38fa7ca820bc3b6e0c00b47da84dfc282ca0ce1e4749c7253d7d3d3ae06b7: Status 404 returned error can't find the container with id 73f38fa7ca820bc3b6e0c00b47da84dfc282ca0ce1e4749c7253d7d3d3ae06b7 Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.589120 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bdd6t"] Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.752776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerStarted","Data":"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284"} Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.752944 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-central-agent" containerID="cri-o://716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" gracePeriod=30 Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.753236 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.753491 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="proxy-httpd" containerID="cri-o://f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" gracePeriod=30 Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.753549 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="sg-core" containerID="cri-o://02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" gracePeriod=30 Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.753586 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-notification-agent" containerID="cri-o://6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" gracePeriod=30 Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.755702 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" event={"ID":"1982e34b-5fe7-4993-88dc-418d7786d001","Type":"ContainerStarted","Data":"73f38fa7ca820bc3b6e0c00b47da84dfc282ca0ce1e4749c7253d7d3d3ae06b7"} Nov 26 06:42:52 crc kubenswrapper[4775]: I1126 06:42:52.773959 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.401023615 podStartE2EDuration="6.773941877s" podCreationTimestamp="2025-11-26 06:42:46 +0000 UTC" firstStartedPulling="2025-11-26 06:42:47.250261203 +0000 UTC m=+1050.611565155" lastFinishedPulling="2025-11-26 06:42:51.623179465 +0000 UTC m=+1054.984483417" observedRunningTime="2025-11-26 06:42:52.773173198 +0000 UTC m=+1056.134477150" watchObservedRunningTime="2025-11-26 06:42:52.773941877 +0000 UTC m=+1056.135245829" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.426922 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-run-httpd\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-sg-core-conf-yaml\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559742 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-config-data\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559800 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z46rx\" (UniqueName: \"kubernetes.io/projected/6ca8e03f-851c-4196-aeb7-f153d60ee295-kube-api-access-z46rx\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559856 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-scripts\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559891 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-combined-ca-bundle\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.559939 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-log-httpd\") pod \"6ca8e03f-851c-4196-aeb7-f153d60ee295\" (UID: \"6ca8e03f-851c-4196-aeb7-f153d60ee295\") " Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.561422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.561586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.566006 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ca8e03f-851c-4196-aeb7-f153d60ee295-kube-api-access-z46rx" (OuterVolumeSpecName: "kube-api-access-z46rx") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "kube-api-access-z46rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.566118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-scripts" (OuterVolumeSpecName: "scripts") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.592347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.657014 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.662986 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z46rx\" (UniqueName: \"kubernetes.io/projected/6ca8e03f-851c-4196-aeb7-f153d60ee295-kube-api-access-z46rx\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.663020 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.663035 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.663049 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.663062 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6ca8e03f-851c-4196-aeb7-f153d60ee295-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.663073 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.696015 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-config-data" (OuterVolumeSpecName: "config-data") pod "6ca8e03f-851c-4196-aeb7-f153d60ee295" (UID: "6ca8e03f-851c-4196-aeb7-f153d60ee295"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.764531 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ca8e03f-851c-4196-aeb7-f153d60ee295-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767211 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" exitCode=0 Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767251 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" exitCode=2 Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767263 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" exitCode=0 Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767273 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" exitCode=0 Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767293 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767294 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerDied","Data":"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284"} Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767406 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerDied","Data":"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5"} Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767417 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerDied","Data":"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8"} Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767427 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerDied","Data":"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504"} Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6ca8e03f-851c-4196-aeb7-f153d60ee295","Type":"ContainerDied","Data":"08602dd948b1ddae5654bcab13bef13a0cc86365102cd5a7766ef40490ec6064"} Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.767452 4775 scope.go:117] "RemoveContainer" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.787838 4775 scope.go:117] "RemoveContainer" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.807532 4775 scope.go:117] "RemoveContainer" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.819936 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.829090 4775 scope.go:117] "RemoveContainer" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.840762 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.851103 4775 scope.go:117] "RemoveContainer" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.852428 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.856264 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": container with ID starting with f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284 not found: ID does not exist" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.856634 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284"} err="failed to get container status \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": rpc error: code = NotFound desc = could not find container \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": container with ID starting with f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.856800 4775 scope.go:117] "RemoveContainer" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.857061 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": container with ID starting with 02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5 not found: ID does not exist" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.857171 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5"} err="failed to get container status \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": rpc error: code = NotFound desc = could not find container \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": container with ID starting with 02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.857255 4775 scope.go:117] "RemoveContainer" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.857754 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="sg-core" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.859913 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="sg-core" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.859997 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="proxy-httpd" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860005 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="proxy-httpd" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.860039 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-central-agent" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860047 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-central-agent" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.860061 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-notification-agent" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860068 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-notification-agent" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860677 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-notification-agent" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860741 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="proxy-httpd" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860759 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="ceilometer-central-agent" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.860768 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" containerName="sg-core" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.863545 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.863668 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.857795 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": container with ID starting with 6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8 not found: ID does not exist" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.864410 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8"} err="failed to get container status \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": rpc error: code = NotFound desc = could not find container \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": container with ID starting with 6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.864438 4775 scope.go:117] "RemoveContainer" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" Nov 26 06:42:53 crc kubenswrapper[4775]: E1126 06:42:53.865725 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": container with ID starting with 716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504 not found: ID does not exist" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.865769 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504"} err="failed to get container status \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": rpc error: code = NotFound desc = could not find container \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": container with ID starting with 716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.865796 4775 scope.go:117] "RemoveContainer" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.866173 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.866326 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.866176 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284"} err="failed to get container status \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": rpc error: code = NotFound desc = could not find container \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": container with ID starting with f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.866505 4775 scope.go:117] "RemoveContainer" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.866828 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5"} err="failed to get container status \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": rpc error: code = NotFound desc = could not find container \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": container with ID starting with 02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.866849 4775 scope.go:117] "RemoveContainer" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.867242 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8"} err="failed to get container status \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": rpc error: code = NotFound desc = could not find container \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": container with ID starting with 6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.867271 4775 scope.go:117] "RemoveContainer" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.867662 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504"} err="failed to get container status \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": rpc error: code = NotFound desc = could not find container \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": container with ID starting with 716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.867695 4775 scope.go:117] "RemoveContainer" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.868021 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284"} err="failed to get container status \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": rpc error: code = NotFound desc = could not find container \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": container with ID starting with f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.868052 4775 scope.go:117] "RemoveContainer" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.868282 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5"} err="failed to get container status \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": rpc error: code = NotFound desc = could not find container \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": container with ID starting with 02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.868313 4775 scope.go:117] "RemoveContainer" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.870167 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8"} err="failed to get container status \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": rpc error: code = NotFound desc = could not find container \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": container with ID starting with 6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.870211 4775 scope.go:117] "RemoveContainer" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.870895 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504"} err="failed to get container status \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": rpc error: code = NotFound desc = could not find container \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": container with ID starting with 716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.870925 4775 scope.go:117] "RemoveContainer" containerID="f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.872142 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284"} err="failed to get container status \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": rpc error: code = NotFound desc = could not find container \"f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284\": container with ID starting with f3a9091c4d6d076c0d1d9c8ad9f20b267b31fb831c3aee27f983d78bfde44284 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.872176 4775 scope.go:117] "RemoveContainer" containerID="02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.873543 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5"} err="failed to get container status \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": rpc error: code = NotFound desc = could not find container \"02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5\": container with ID starting with 02360f2ac93c27a2ada53853af6e239c1e284c58c43da16c6f5b4609e14376b5 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.873594 4775 scope.go:117] "RemoveContainer" containerID="6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.876062 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8"} err="failed to get container status \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": rpc error: code = NotFound desc = could not find container \"6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8\": container with ID starting with 6dd091eca766428349dde138cdef6d3986dda86ffcb8de6fc6a175bf015cc5d8 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.876133 4775 scope.go:117] "RemoveContainer" containerID="716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.876396 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504"} err="failed to get container status \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": rpc error: code = NotFound desc = could not find container \"716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504\": container with ID starting with 716b8d9f1397ae5682996c93665b55ba9e41ae912ded214e0f7dba03ceafc504 not found: ID does not exist" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.967886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-scripts\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.968016 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-config-data\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.968247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-run-httpd\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.968298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lw5\" (UniqueName: \"kubernetes.io/projected/e3dd9226-b058-4502-92f4-85b9fcc99ef1-kube-api-access-t8lw5\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.968326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.968491 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-log-httpd\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:53 crc kubenswrapper[4775]: I1126 06:42:53.968542 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.069911 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-log-httpd\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.069950 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.070020 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-scripts\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.070054 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-config-data\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.070107 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lw5\" (UniqueName: \"kubernetes.io/projected/e3dd9226-b058-4502-92f4-85b9fcc99ef1-kube-api-access-t8lw5\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.070122 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-run-httpd\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.070138 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.071297 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-run-httpd\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.071380 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-log-httpd\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.075205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-scripts\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.075562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.076019 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-config-data\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.076632 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.092629 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lw5\" (UniqueName: \"kubernetes.io/projected/e3dd9226-b058-4502-92f4-85b9fcc99ef1-kube-api-access-t8lw5\") pod \"ceilometer-0\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.181365 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:42:54 crc kubenswrapper[4775]: E1126 06:42:54.339916 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb607340_2aa9_4b57_a4b7_a6a64457cae3.slice/crio-conmon-aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb607340_2aa9_4b57_a4b7_a6a64457cae3.slice/crio-aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0.scope\": RecentStats: unable to find data in memory cache]" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.621981 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.691923 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-config-data\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.692009 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-secret-key\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.692034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-combined-ca-bundle\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.692063 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb607340-2aa9-4b57-a4b7-a6a64457cae3-logs\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.692084 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-tls-certs\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.692109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vltpp\" (UniqueName: \"kubernetes.io/projected/fb607340-2aa9-4b57-a4b7-a6a64457cae3-kube-api-access-vltpp\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.692135 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-scripts\") pod \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\" (UID: \"fb607340-2aa9-4b57-a4b7-a6a64457cae3\") " Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.693004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb607340-2aa9-4b57-a4b7-a6a64457cae3-logs" (OuterVolumeSpecName: "logs") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.700905 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.701536 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb607340-2aa9-4b57-a4b7-a6a64457cae3-kube-api-access-vltpp" (OuterVolumeSpecName: "kube-api-access-vltpp") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "kube-api-access-vltpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.731643 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-config-data" (OuterVolumeSpecName: "config-data") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.773422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.776881 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.781310 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-scripts" (OuterVolumeSpecName: "scripts") pod "fb607340-2aa9-4b57-a4b7-a6a64457cae3" (UID: "fb607340-2aa9-4b57-a4b7-a6a64457cae3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796876 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb607340-2aa9-4b57-a4b7-a6a64457cae3-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796914 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796929 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vltpp\" (UniqueName: \"kubernetes.io/projected/fb607340-2aa9-4b57-a4b7-a6a64457cae3-kube-api-access-vltpp\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796943 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796954 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb607340-2aa9-4b57-a4b7-a6a64457cae3-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796966 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.796977 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb607340-2aa9-4b57-a4b7-a6a64457cae3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.805768 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerID="aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0" exitCode=137 Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.805852 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5886fff4d-zcj7s" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.805869 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5886fff4d-zcj7s" event={"ID":"fb607340-2aa9-4b57-a4b7-a6a64457cae3","Type":"ContainerDied","Data":"aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0"} Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.805897 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5886fff4d-zcj7s" event={"ID":"fb607340-2aa9-4b57-a4b7-a6a64457cae3","Type":"ContainerDied","Data":"13cab573fa3085fbc1a2ee95bad6d2f475a0e4cdf60ff95d53d3e4585bc1541c"} Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.805917 4775 scope.go:117] "RemoveContainer" containerID="4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a" Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.835507 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.903291 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5886fff4d-zcj7s"] Nov 26 06:42:54 crc kubenswrapper[4775]: I1126 06:42:54.912551 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5886fff4d-zcj7s"] Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.070436 4775 scope.go:117] "RemoveContainer" containerID="aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.099990 4775 scope.go:117] "RemoveContainer" containerID="4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a" Nov 26 06:42:55 crc kubenswrapper[4775]: E1126 06:42:55.100333 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a\": container with ID starting with 4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a not found: ID does not exist" containerID="4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.100375 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a"} err="failed to get container status \"4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a\": rpc error: code = NotFound desc = could not find container \"4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a\": container with ID starting with 4fe108aefb447e707b0264f3d9baf8d2c5b96fb5bd704f1bcabf9515d09e271a not found: ID does not exist" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.100402 4775 scope.go:117] "RemoveContainer" containerID="aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0" Nov 26 06:42:55 crc kubenswrapper[4775]: E1126 06:42:55.100689 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0\": container with ID starting with aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0 not found: ID does not exist" containerID="aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.100728 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0"} err="failed to get container status \"aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0\": rpc error: code = NotFound desc = could not find container \"aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0\": container with ID starting with aee8b41fdceb41ad507e690444b78ac6467fa1ae41ffc67f0580fd15340316a0 not found: ID does not exist" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.367526 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ca8e03f-851c-4196-aeb7-f153d60ee295" path="/var/lib/kubelet/pods/6ca8e03f-851c-4196-aeb7-f153d60ee295/volumes" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.370161 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" path="/var/lib/kubelet/pods/fb607340-2aa9-4b57-a4b7-a6a64457cae3/volumes" Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.683576 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.825141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerStarted","Data":"fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d"} Nov 26 06:42:55 crc kubenswrapper[4775]: I1126 06:42:55.825377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerStarted","Data":"3072aa0171e87bf893852c31f30624881fe788fc5e1f8f970cd12be4cf9a8e5d"} Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.431151 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.431190 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.477983 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.505461 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.836635 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerStarted","Data":"0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a"} Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.836818 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:56 crc kubenswrapper[4775]: I1126 06:42:56.836833 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.042070 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.042114 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.077571 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.096940 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.846469 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerStarted","Data":"f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030"} Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.846806 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 26 06:42:57 crc kubenswrapper[4775]: I1126 06:42:57.846822 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 26 06:42:58 crc kubenswrapper[4775]: I1126 06:42:58.678023 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 26 06:42:58 crc kubenswrapper[4775]: I1126 06:42:58.856109 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:42:59 crc kubenswrapper[4775]: I1126 06:42:59.171926 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 26 06:43:00 crc kubenswrapper[4775]: I1126 06:43:00.099865 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 26 06:43:00 crc kubenswrapper[4775]: I1126 06:43:00.099990 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 06:43:00 crc kubenswrapper[4775]: I1126 06:43:00.114176 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.908543 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerStarted","Data":"58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2"} Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.909425 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-central-agent" containerID="cri-o://fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d" gracePeriod=30 Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.909845 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.910221 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="proxy-httpd" containerID="cri-o://58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2" gracePeriod=30 Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.910317 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="sg-core" containerID="cri-o://f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030" gracePeriod=30 Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.910390 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-notification-agent" containerID="cri-o://0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a" gracePeriod=30 Nov 26 06:43:03 crc kubenswrapper[4775]: I1126 06:43:03.912354 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" event={"ID":"1982e34b-5fe7-4993-88dc-418d7786d001","Type":"ContainerStarted","Data":"c90cceebc5643f4620e78dac89d9b6efaf35537fe59fd1f1bbb82f1144e5d1fb"} Nov 26 06:43:04 crc kubenswrapper[4775]: I1126 06:43:04.270756 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.135634278 podStartE2EDuration="11.270734617s" podCreationTimestamp="2025-11-26 06:42:53 +0000 UTC" firstStartedPulling="2025-11-26 06:42:54.84403205 +0000 UTC m=+1058.205335992" lastFinishedPulling="2025-11-26 06:43:02.979132379 +0000 UTC m=+1066.340436331" observedRunningTime="2025-11-26 06:43:04.261195867 +0000 UTC m=+1067.622499829" watchObservedRunningTime="2025-11-26 06:43:04.270734617 +0000 UTC m=+1067.632038569" Nov 26 06:43:04 crc kubenswrapper[4775]: I1126 06:43:04.286820 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" podStartSLOduration=2.879325809 podStartE2EDuration="13.28679954s" podCreationTimestamp="2025-11-26 06:42:51 +0000 UTC" firstStartedPulling="2025-11-26 06:42:52.58845517 +0000 UTC m=+1055.949759122" lastFinishedPulling="2025-11-26 06:43:02.995928901 +0000 UTC m=+1066.357232853" observedRunningTime="2025-11-26 06:43:04.276864001 +0000 UTC m=+1067.638167963" watchObservedRunningTime="2025-11-26 06:43:04.28679954 +0000 UTC m=+1067.648103502" Nov 26 06:43:04 crc kubenswrapper[4775]: I1126 06:43:04.923928 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerID="f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030" exitCode=2 Nov 26 06:43:04 crc kubenswrapper[4775]: I1126 06:43:04.923963 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerID="0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a" exitCode=0 Nov 26 06:43:04 crc kubenswrapper[4775]: I1126 06:43:04.923997 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerDied","Data":"f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030"} Nov 26 06:43:04 crc kubenswrapper[4775]: I1126 06:43:04.924055 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerDied","Data":"0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a"} Nov 26 06:43:05 crc kubenswrapper[4775]: I1126 06:43:05.938825 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerID="58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2" exitCode=0 Nov 26 06:43:05 crc kubenswrapper[4775]: I1126 06:43:05.938893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerDied","Data":"58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2"} Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.575601 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.623973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-run-httpd\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624087 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-sg-core-conf-yaml\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624193 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-config-data\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624238 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-combined-ca-bundle\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624278 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8lw5\" (UniqueName: \"kubernetes.io/projected/e3dd9226-b058-4502-92f4-85b9fcc99ef1-kube-api-access-t8lw5\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624312 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-log-httpd\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624354 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-scripts\") pod \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\" (UID: \"e3dd9226-b058-4502-92f4-85b9fcc99ef1\") " Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.624821 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.625329 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.636009 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3dd9226-b058-4502-92f4-85b9fcc99ef1-kube-api-access-t8lw5" (OuterVolumeSpecName: "kube-api-access-t8lw5") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "kube-api-access-t8lw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.636085 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-scripts" (OuterVolumeSpecName: "scripts") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.659000 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.706056 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.720134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-config-data" (OuterVolumeSpecName: "config-data") pod "e3dd9226-b058-4502-92f4-85b9fcc99ef1" (UID: "e3dd9226-b058-4502-92f4-85b9fcc99ef1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727339 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727369 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727524 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8lw5\" (UniqueName: \"kubernetes.io/projected/e3dd9226-b058-4502-92f4-85b9fcc99ef1-kube-api-access-t8lw5\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727540 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727551 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727559 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e3dd9226-b058-4502-92f4-85b9fcc99ef1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.727569 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e3dd9226-b058-4502-92f4-85b9fcc99ef1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.951341 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerID="fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d" exitCode=0 Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.951393 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerDied","Data":"fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d"} Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.951422 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e3dd9226-b058-4502-92f4-85b9fcc99ef1","Type":"ContainerDied","Data":"3072aa0171e87bf893852c31f30624881fe788fc5e1f8f970cd12be4cf9a8e5d"} Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.951442 4775 scope.go:117] "RemoveContainer" containerID="58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.951464 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:06 crc kubenswrapper[4775]: I1126 06:43:06.985919 4775 scope.go:117] "RemoveContainer" containerID="f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.023356 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.023369 4775 scope.go:117] "RemoveContainer" containerID="0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.036598 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.045865 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.046345 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-notification-agent" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046367 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-notification-agent" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.046385 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-central-agent" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046394 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-central-agent" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.046411 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046420 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.046433 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon-log" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046441 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon-log" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.046464 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="sg-core" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046490 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="sg-core" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.046521 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="proxy-httpd" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046529 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="proxy-httpd" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046780 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="sg-core" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046805 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-notification-agent" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046817 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="ceilometer-central-agent" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046831 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046840 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb607340-2aa9-4b57-a4b7-a6a64457cae3" containerName="horizon-log" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.046859 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" containerName="proxy-httpd" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.048877 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.055240 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.057979 4775 scope.go:117] "RemoveContainer" containerID="fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.058248 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.059059 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.078932 4775 scope.go:117] "RemoveContainer" containerID="58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.080280 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2\": container with ID starting with 58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2 not found: ID does not exist" containerID="58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.080314 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2"} err="failed to get container status \"58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2\": rpc error: code = NotFound desc = could not find container \"58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2\": container with ID starting with 58a3c5e1eac0b84b6f076b0e937a01c65a10d450fd541627fd41cca6e24ad5b2 not found: ID does not exist" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.080335 4775 scope.go:117] "RemoveContainer" containerID="f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.080632 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030\": container with ID starting with f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030 not found: ID does not exist" containerID="f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.080674 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030"} err="failed to get container status \"f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030\": rpc error: code = NotFound desc = could not find container \"f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030\": container with ID starting with f99a2316bb3ba94130f220ae6a91e68b9e59569ea7d12afd60c57fd82760d030 not found: ID does not exist" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.080701 4775 scope.go:117] "RemoveContainer" containerID="0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.081080 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a\": container with ID starting with 0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a not found: ID does not exist" containerID="0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.081105 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a"} err="failed to get container status \"0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a\": rpc error: code = NotFound desc = could not find container \"0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a\": container with ID starting with 0cf74835986090c32fbc8059b7b358c16b6adb5ba93d64829e377be077cf667a not found: ID does not exist" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.081120 4775 scope.go:117] "RemoveContainer" containerID="fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d" Nov 26 06:43:07 crc kubenswrapper[4775]: E1126 06:43:07.081307 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d\": container with ID starting with fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d not found: ID does not exist" containerID="fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.081324 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d"} err="failed to get container status \"fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d\": rpc error: code = NotFound desc = could not find container \"fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d\": container with ID starting with fa7dddad0dc66cdbac3e1aac9dfceb9c4398dcef3c3fa2c876197db73b05df6d not found: ID does not exist" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.133839 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47xzg\" (UniqueName: \"kubernetes.io/projected/678b727b-e044-42ea-97b0-3963fbcc32fc-kube-api-access-47xzg\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.133916 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-run-httpd\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.133952 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-scripts\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.133994 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-config-data\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.134023 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.134066 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-log-httpd\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.134080 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235705 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-run-httpd\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-scripts\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235836 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-config-data\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235871 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-log-httpd\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.235977 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47xzg\" (UniqueName: \"kubernetes.io/projected/678b727b-e044-42ea-97b0-3963fbcc32fc-kube-api-access-47xzg\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.236553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-log-httpd\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.237027 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-run-httpd\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.240692 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.240751 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.241009 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-config-data\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.241421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-scripts\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.254128 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47xzg\" (UniqueName: \"kubernetes.io/projected/678b727b-e044-42ea-97b0-3963fbcc32fc-kube-api-access-47xzg\") pod \"ceilometer-0\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.339989 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3dd9226-b058-4502-92f4-85b9fcc99ef1" path="/var/lib/kubelet/pods/e3dd9226-b058-4502-92f4-85b9fcc99ef1/volumes" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.375188 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.824610 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:07 crc kubenswrapper[4775]: I1126 06:43:07.963130 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerStarted","Data":"52d2160739fe51b7d63ab12e7ca0481556d414c793347fb307228893d7b08333"} Nov 26 06:43:08 crc kubenswrapper[4775]: I1126 06:43:08.977968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerStarted","Data":"aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8"} Nov 26 06:43:09 crc kubenswrapper[4775]: I1126 06:43:09.990873 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerStarted","Data":"8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d"} Nov 26 06:43:09 crc kubenswrapper[4775]: I1126 06:43:09.991182 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerStarted","Data":"21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609"} Nov 26 06:43:12 crc kubenswrapper[4775]: I1126 06:43:12.010626 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerStarted","Data":"6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9"} Nov 26 06:43:12 crc kubenswrapper[4775]: I1126 06:43:12.010869 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:43:12 crc kubenswrapper[4775]: I1126 06:43:12.047338 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.072237698 podStartE2EDuration="6.047314854s" podCreationTimestamp="2025-11-26 06:43:06 +0000 UTC" firstStartedPulling="2025-11-26 06:43:07.831650881 +0000 UTC m=+1071.192954833" lastFinishedPulling="2025-11-26 06:43:10.806728037 +0000 UTC m=+1074.168031989" observedRunningTime="2025-11-26 06:43:12.039525239 +0000 UTC m=+1075.400829201" watchObservedRunningTime="2025-11-26 06:43:12.047314854 +0000 UTC m=+1075.408618806" Nov 26 06:43:13 crc kubenswrapper[4775]: I1126 06:43:13.844083 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:14 crc kubenswrapper[4775]: I1126 06:43:14.029841 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-central-agent" containerID="cri-o://aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8" gracePeriod=30 Nov 26 06:43:14 crc kubenswrapper[4775]: I1126 06:43:14.029937 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="sg-core" containerID="cri-o://8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d" gracePeriod=30 Nov 26 06:43:14 crc kubenswrapper[4775]: I1126 06:43:14.029986 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-notification-agent" containerID="cri-o://21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609" gracePeriod=30 Nov 26 06:43:14 crc kubenswrapper[4775]: I1126 06:43:14.030083 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="proxy-httpd" containerID="cri-o://6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9" gracePeriod=30 Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.039689 4775 generic.go:334] "Generic (PLEG): container finished" podID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerID="6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9" exitCode=0 Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.040289 4775 generic.go:334] "Generic (PLEG): container finished" podID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerID="8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d" exitCode=2 Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.040297 4775 generic.go:334] "Generic (PLEG): container finished" podID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerID="21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609" exitCode=0 Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.039894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerDied","Data":"6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9"} Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.040326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerDied","Data":"8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d"} Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.040336 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerDied","Data":"21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609"} Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.435958 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492324 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-scripts\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-config-data\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492469 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-log-httpd\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-sg-core-conf-yaml\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-combined-ca-bundle\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47xzg\" (UniqueName: \"kubernetes.io/projected/678b727b-e044-42ea-97b0-3963fbcc32fc-kube-api-access-47xzg\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.492705 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-run-httpd\") pod \"678b727b-e044-42ea-97b0-3963fbcc32fc\" (UID: \"678b727b-e044-42ea-97b0-3963fbcc32fc\") " Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.493353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.499157 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.501278 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/678b727b-e044-42ea-97b0-3963fbcc32fc-kube-api-access-47xzg" (OuterVolumeSpecName: "kube-api-access-47xzg") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "kube-api-access-47xzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.522623 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-scripts" (OuterVolumeSpecName: "scripts") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.535222 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.566853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.588425 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-config-data" (OuterVolumeSpecName: "config-data") pod "678b727b-e044-42ea-97b0-3963fbcc32fc" (UID: "678b727b-e044-42ea-97b0-3963fbcc32fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600125 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600333 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47xzg\" (UniqueName: \"kubernetes.io/projected/678b727b-e044-42ea-97b0-3963fbcc32fc-kube-api-access-47xzg\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600399 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600458 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600513 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600606 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/678b727b-e044-42ea-97b0-3963fbcc32fc-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:15 crc kubenswrapper[4775]: I1126 06:43:15.600666 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/678b727b-e044-42ea-97b0-3963fbcc32fc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.052478 4775 generic.go:334] "Generic (PLEG): container finished" podID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerID="aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8" exitCode=0 Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.052526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerDied","Data":"aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8"} Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.052567 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"678b727b-e044-42ea-97b0-3963fbcc32fc","Type":"ContainerDied","Data":"52d2160739fe51b7d63ab12e7ca0481556d414c793347fb307228893d7b08333"} Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.052587 4775 scope.go:117] "RemoveContainer" containerID="6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.054475 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.091335 4775 scope.go:117] "RemoveContainer" containerID="8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.121006 4775 scope.go:117] "RemoveContainer" containerID="21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.125827 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.146690 4775 scope.go:117] "RemoveContainer" containerID="aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.163627 4775 scope.go:117] "RemoveContainer" containerID="6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.164144 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9\": container with ID starting with 6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9 not found: ID does not exist" containerID="6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.164281 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9"} err="failed to get container status \"6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9\": rpc error: code = NotFound desc = could not find container \"6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9\": container with ID starting with 6dd8f551a00a9557198abf2b434f69da113914daa4583f2e6cc5f569a1a0a9b9 not found: ID does not exist" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.164349 4775 scope.go:117] "RemoveContainer" containerID="8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.164683 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d\": container with ID starting with 8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d not found: ID does not exist" containerID="8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.164743 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d"} err="failed to get container status \"8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d\": rpc error: code = NotFound desc = could not find container \"8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d\": container with ID starting with 8fec4720372932a002275168e4632b3b73a8b0f24422c4be10e6c9147f371e7d not found: ID does not exist" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.164767 4775 scope.go:117] "RemoveContainer" containerID="21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.165035 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609\": container with ID starting with 21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609 not found: ID does not exist" containerID="21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.165159 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609"} err="failed to get container status \"21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609\": rpc error: code = NotFound desc = could not find container \"21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609\": container with ID starting with 21f8e801429cf58961583fcf33ad04be7174cd93b8409619f2ebf305fb6ed609 not found: ID does not exist" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.165256 4775 scope.go:117] "RemoveContainer" containerID="aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.165708 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8\": container with ID starting with aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8 not found: ID does not exist" containerID="aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.165807 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8"} err="failed to get container status \"aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8\": rpc error: code = NotFound desc = could not find container \"aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8\": container with ID starting with aa4a396fb949aa4629f037322e971ddd041883bf133468dae763f09e97c8e0f8 not found: ID does not exist" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.171616 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.183508 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.184284 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-central-agent" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184307 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-central-agent" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.184325 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-notification-agent" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184332 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-notification-agent" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.184350 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="sg-core" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184356 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="sg-core" Nov 26 06:43:16 crc kubenswrapper[4775]: E1126 06:43:16.184370 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="proxy-httpd" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184376 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="proxy-httpd" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184852 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="proxy-httpd" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184874 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-central-agent" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184895 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="sg-core" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.184903 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" containerName="ceilometer-notification-agent" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.187732 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.195133 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.201232 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.201593 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213106 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-log-httpd\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-run-httpd\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213336 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-scripts\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213360 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-config-data\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.213539 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94xzh\" (UniqueName: \"kubernetes.io/projected/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-kube-api-access-94xzh\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.314790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-log-httpd\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.314854 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.314879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-run-httpd\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.314908 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-scripts\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.314930 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.314947 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-config-data\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.315015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94xzh\" (UniqueName: \"kubernetes.io/projected/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-kube-api-access-94xzh\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.315552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-log-httpd\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.315818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-run-httpd\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.318963 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.319660 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-scripts\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.324774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-config-data\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.339002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.344824 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94xzh\" (UniqueName: \"kubernetes.io/projected/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-kube-api-access-94xzh\") pod \"ceilometer-0\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " pod="openstack/ceilometer-0" Nov 26 06:43:16 crc kubenswrapper[4775]: I1126 06:43:16.509443 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:17 crc kubenswrapper[4775]: I1126 06:43:17.000237 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:17 crc kubenswrapper[4775]: I1126 06:43:17.068424 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerStarted","Data":"e2253269920d3c5f264f18afc2f71d78ad7e6f83da65839dc69e741750b90c1a"} Nov 26 06:43:17 crc kubenswrapper[4775]: I1126 06:43:17.071874 4775 generic.go:334] "Generic (PLEG): container finished" podID="1982e34b-5fe7-4993-88dc-418d7786d001" containerID="c90cceebc5643f4620e78dac89d9b6efaf35537fe59fd1f1bbb82f1144e5d1fb" exitCode=0 Nov 26 06:43:17 crc kubenswrapper[4775]: I1126 06:43:17.071934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" event={"ID":"1982e34b-5fe7-4993-88dc-418d7786d001","Type":"ContainerDied","Data":"c90cceebc5643f4620e78dac89d9b6efaf35537fe59fd1f1bbb82f1144e5d1fb"} Nov 26 06:43:17 crc kubenswrapper[4775]: I1126 06:43:17.341695 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="678b727b-e044-42ea-97b0-3963fbcc32fc" path="/var/lib/kubelet/pods/678b727b-e044-42ea-97b0-3963fbcc32fc/volumes" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.105198 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerStarted","Data":"d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811"} Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.422091 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.562133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-combined-ca-bundle\") pod \"1982e34b-5fe7-4993-88dc-418d7786d001\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.562168 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nzj8\" (UniqueName: \"kubernetes.io/projected/1982e34b-5fe7-4993-88dc-418d7786d001-kube-api-access-2nzj8\") pod \"1982e34b-5fe7-4993-88dc-418d7786d001\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.562233 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-scripts\") pod \"1982e34b-5fe7-4993-88dc-418d7786d001\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.563191 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-config-data\") pod \"1982e34b-5fe7-4993-88dc-418d7786d001\" (UID: \"1982e34b-5fe7-4993-88dc-418d7786d001\") " Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.566827 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1982e34b-5fe7-4993-88dc-418d7786d001-kube-api-access-2nzj8" (OuterVolumeSpecName: "kube-api-access-2nzj8") pod "1982e34b-5fe7-4993-88dc-418d7786d001" (UID: "1982e34b-5fe7-4993-88dc-418d7786d001"). InnerVolumeSpecName "kube-api-access-2nzj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.579133 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-scripts" (OuterVolumeSpecName: "scripts") pod "1982e34b-5fe7-4993-88dc-418d7786d001" (UID: "1982e34b-5fe7-4993-88dc-418d7786d001"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.587275 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1982e34b-5fe7-4993-88dc-418d7786d001" (UID: "1982e34b-5fe7-4993-88dc-418d7786d001"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.609270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-config-data" (OuterVolumeSpecName: "config-data") pod "1982e34b-5fe7-4993-88dc-418d7786d001" (UID: "1982e34b-5fe7-4993-88dc-418d7786d001"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.665321 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.665355 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nzj8\" (UniqueName: \"kubernetes.io/projected/1982e34b-5fe7-4993-88dc-418d7786d001-kube-api-access-2nzj8\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.665366 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:18 crc kubenswrapper[4775]: I1126 06:43:18.665378 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1982e34b-5fe7-4993-88dc-418d7786d001-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.126358 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerStarted","Data":"4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c"} Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.126573 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerStarted","Data":"fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7"} Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.128321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" event={"ID":"1982e34b-5fe7-4993-88dc-418d7786d001","Type":"ContainerDied","Data":"73f38fa7ca820bc3b6e0c00b47da84dfc282ca0ce1e4749c7253d7d3d3ae06b7"} Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.128357 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73f38fa7ca820bc3b6e0c00b47da84dfc282ca0ce1e4749c7253d7d3d3ae06b7" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.128434 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bdd6t" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.217460 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 26 06:43:19 crc kubenswrapper[4775]: E1126 06:43:19.218165 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1982e34b-5fe7-4993-88dc-418d7786d001" containerName="nova-cell0-conductor-db-sync" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.218187 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1982e34b-5fe7-4993-88dc-418d7786d001" containerName="nova-cell0-conductor-db-sync" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.218377 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1982e34b-5fe7-4993-88dc-418d7786d001" containerName="nova-cell0-conductor-db-sync" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.236501 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.236599 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.238748 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.238835 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6bjxj" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.376409 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.376476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqttx\" (UniqueName: \"kubernetes.io/projected/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-kube-api-access-vqttx\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.376571 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.478267 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.478662 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqttx\" (UniqueName: \"kubernetes.io/projected/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-kube-api-access-vqttx\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.478828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.490886 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.491160 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.599525 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqttx\" (UniqueName: \"kubernetes.io/projected/76cb0fc1-4a71-4ae4-9b2e-7483efe206d7-kube-api-access-vqttx\") pod \"nova-cell0-conductor-0\" (UID: \"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7\") " pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:19 crc kubenswrapper[4775]: I1126 06:43:19.860510 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:20 crc kubenswrapper[4775]: I1126 06:43:20.373809 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 26 06:43:20 crc kubenswrapper[4775]: W1126 06:43:20.383360 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76cb0fc1_4a71_4ae4_9b2e_7483efe206d7.slice/crio-362969f1dda52e36f4a21bfdf8e1e1408ba6439085afceff6b3dc30bab81977f WatchSource:0}: Error finding container 362969f1dda52e36f4a21bfdf8e1e1408ba6439085afceff6b3dc30bab81977f: Status 404 returned error can't find the container with id 362969f1dda52e36f4a21bfdf8e1e1408ba6439085afceff6b3dc30bab81977f Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.155876 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7","Type":"ContainerStarted","Data":"508db1df6878d41f91fa476ff4b6af380d2fda6715faeb1b63b39d3703aaafdd"} Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.155925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"76cb0fc1-4a71-4ae4-9b2e-7483efe206d7","Type":"ContainerStarted","Data":"362969f1dda52e36f4a21bfdf8e1e1408ba6439085afceff6b3dc30bab81977f"} Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.156031 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.159555 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerStarted","Data":"8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf"} Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.159993 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.184989 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.184962184 podStartE2EDuration="2.184962184s" podCreationTimestamp="2025-11-26 06:43:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:21.180675197 +0000 UTC m=+1084.541979179" watchObservedRunningTime="2025-11-26 06:43:21.184962184 +0000 UTC m=+1084.546266166" Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.420122 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:43:21 crc kubenswrapper[4775]: I1126 06:43:21.422059 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:43:29 crc kubenswrapper[4775]: I1126 06:43:29.916242 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 26 06:43:29 crc kubenswrapper[4775]: I1126 06:43:29.948063 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=10.694864561 podStartE2EDuration="13.948035318s" podCreationTimestamp="2025-11-26 06:43:16 +0000 UTC" firstStartedPulling="2025-11-26 06:43:16.986222926 +0000 UTC m=+1080.347526888" lastFinishedPulling="2025-11-26 06:43:20.239393693 +0000 UTC m=+1083.600697645" observedRunningTime="2025-11-26 06:43:21.218675721 +0000 UTC m=+1084.579979693" watchObservedRunningTime="2025-11-26 06:43:29.948035318 +0000 UTC m=+1093.309339300" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.424007 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xg4dj"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.425607 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.427766 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.428044 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.448099 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xg4dj"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.532572 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-config-data\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.532825 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8njqd\" (UniqueName: \"kubernetes.io/projected/59a2dfdf-8626-45d9-a37b-91272ecbc570-kube-api-access-8njqd\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.532971 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.533239 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-scripts\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.632224 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.634580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8njqd\" (UniqueName: \"kubernetes.io/projected/59a2dfdf-8626-45d9-a37b-91272ecbc570-kube-api-access-8njqd\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.634627 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.634683 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-scripts\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.634736 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-config-data\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.635595 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.640051 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.640657 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-scripts\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.646491 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-config-data\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.653851 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.658000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.686280 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8njqd\" (UniqueName: \"kubernetes.io/projected/59a2dfdf-8626-45d9-a37b-91272ecbc570-kube-api-access-8njqd\") pod \"nova-cell0-cell-mapping-xg4dj\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.689798 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.691692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.700179 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.714846 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.738434 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.738689 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jm94\" (UniqueName: \"kubernetes.io/projected/7119330c-76fa-4850-9372-f3dd55c20247-kube-api-access-9jm94\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.738863 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-config-data\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.738937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7119330c-76fa-4850-9372-f3dd55c20247-logs\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.778495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863763 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3529d7-ae55-48f8-9c32-392432471665-logs\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863823 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jm94\" (UniqueName: \"kubernetes.io/projected/7119330c-76fa-4850-9372-f3dd55c20247-kube-api-access-9jm94\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863891 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs4pb\" (UniqueName: \"kubernetes.io/projected/6a3529d7-ae55-48f8-9c32-392432471665-kube-api-access-vs4pb\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863927 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-config-data\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863957 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-config-data\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.863973 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7119330c-76fa-4850-9372-f3dd55c20247-logs\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.864390 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7119330c-76fa-4850-9372-f3dd55c20247-logs\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.868050 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.869242 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.872987 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.875680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.891887 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-config-data\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.901682 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-7q5vm"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.903004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jm94\" (UniqueName: \"kubernetes.io/projected/7119330c-76fa-4850-9372-f3dd55c20247-kube-api-access-9jm94\") pod \"nova-api-0\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " pod="openstack/nova-api-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.903808 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.941008 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.956371 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-7q5vm"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.966409 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.966476 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs4pb\" (UniqueName: \"kubernetes.io/projected/6a3529d7-ae55-48f8-9c32-392432471665-kube-api-access-vs4pb\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.966516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-config-data\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.966582 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3529d7-ae55-48f8-9c32-392432471665-logs\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.966991 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3529d7-ae55-48f8-9c32-392432471665-logs\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.971873 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-config-data\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.971928 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.978655 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.979876 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.984340 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.984633 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs4pb\" (UniqueName: \"kubernetes.io/projected/6a3529d7-ae55-48f8-9c32-392432471665-kube-api-access-vs4pb\") pod \"nova-metadata-0\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " pod="openstack/nova-metadata-0" Nov 26 06:43:30 crc kubenswrapper[4775]: I1126 06:43:30.997594 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.067893 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-config-data\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.067988 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4q4n\" (UniqueName: \"kubernetes.io/projected/1418942b-2d58-49c9-b24f-9bb582ea3ed8-kube-api-access-v4q4n\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068049 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcpzf\" (UniqueName: \"kubernetes.io/projected/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-kube-api-access-lcpzf\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068071 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068098 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068138 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068169 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068198 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-config\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.068262 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.078505 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.121460 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169699 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169756 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-config\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-config-data\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169944 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169976 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nf42\" (UniqueName: \"kubernetes.io/projected/dd0776fb-8edf-4508-98c3-5267fb06cd56-kube-api-access-7nf42\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.169996 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4q4n\" (UniqueName: \"kubernetes.io/projected/1418942b-2d58-49c9-b24f-9bb582ea3ed8-kube-api-access-v4q4n\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.170039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcpzf\" (UniqueName: \"kubernetes.io/projected/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-kube-api-access-lcpzf\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.170658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.171760 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.171789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.172115 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-config\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.172645 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.173591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.175588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-config-data\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.186951 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcpzf\" (UniqueName: \"kubernetes.io/projected/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-kube-api-access-lcpzf\") pod \"dnsmasq-dns-845d6d6f59-7q5vm\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.188888 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4q4n\" (UniqueName: \"kubernetes.io/projected/1418942b-2d58-49c9-b24f-9bb582ea3ed8-kube-api-access-v4q4n\") pod \"nova-scheduler-0\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.271631 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nf42\" (UniqueName: \"kubernetes.io/projected/dd0776fb-8edf-4508-98c3-5267fb06cd56-kube-api-access-7nf42\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.271819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.271850 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.276178 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.276421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.302218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nf42\" (UniqueName: \"kubernetes.io/projected/dd0776fb-8edf-4508-98c3-5267fb06cd56-kube-api-access-7nf42\") pod \"nova-cell1-novncproxy-0\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.302663 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.325770 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.336007 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.399239 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xg4dj"] Nov 26 06:43:31 crc kubenswrapper[4775]: W1126 06:43:31.445991 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59a2dfdf_8626_45d9_a37b_91272ecbc570.slice/crio-a8974ea658ab00fda5797f8ea8d0a13fd60c609b11c7e075c75a52b28c443b5d WatchSource:0}: Error finding container a8974ea658ab00fda5797f8ea8d0a13fd60c609b11c7e075c75a52b28c443b5d: Status 404 returned error can't find the container with id a8974ea658ab00fda5797f8ea8d0a13fd60c609b11c7e075c75a52b28c443b5d Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.516484 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6prq8"] Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.517785 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.520079 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.520252 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.528901 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6prq8"] Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.588945 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.681698 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.683747 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5m8b\" (UniqueName: \"kubernetes.io/projected/5c567945-b393-4d26-8916-cb24e2c38e07-kube-api-access-g5m8b\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.683803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.683859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-config-data\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.683919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-scripts\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.707461 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-7q5vm"] Nov 26 06:43:31 crc kubenswrapper[4775]: W1126 06:43:31.727934 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef02f1ef_024a_4c85_826f_7b0c4749dbdd.slice/crio-512fdbe0f2201de777a9ed7a34e2d5d03f0e05267cce70455224d87aded20077 WatchSource:0}: Error finding container 512fdbe0f2201de777a9ed7a34e2d5d03f0e05267cce70455224d87aded20077: Status 404 returned error can't find the container with id 512fdbe0f2201de777a9ed7a34e2d5d03f0e05267cce70455224d87aded20077 Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.785719 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5m8b\" (UniqueName: \"kubernetes.io/projected/5c567945-b393-4d26-8916-cb24e2c38e07-kube-api-access-g5m8b\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.786193 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.786291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-config-data\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.786387 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-scripts\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.789739 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-scripts\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.792263 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-config-data\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.793596 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.805169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5m8b\" (UniqueName: \"kubernetes.io/projected/5c567945-b393-4d26-8916-cb24e2c38e07-kube-api-access-g5m8b\") pod \"nova-cell1-conductor-db-sync-6prq8\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.825174 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:31 crc kubenswrapper[4775]: W1126 06:43:31.840296 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1418942b_2d58_49c9_b24f_9bb582ea3ed8.slice/crio-30954ef2d49e869b1da789bea2bcbf118fb6cc4de22e028736e2cdb993bb2e58 WatchSource:0}: Error finding container 30954ef2d49e869b1da789bea2bcbf118fb6cc4de22e028736e2cdb993bb2e58: Status 404 returned error can't find the container with id 30954ef2d49e869b1da789bea2bcbf118fb6cc4de22e028736e2cdb993bb2e58 Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.851495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:31 crc kubenswrapper[4775]: I1126 06:43:31.957588 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.300994 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6prq8"] Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.312005 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xg4dj" event={"ID":"59a2dfdf-8626-45d9-a37b-91272ecbc570","Type":"ContainerStarted","Data":"5cf4f039200a4d89cb81f8faf7b4ffb773a750ebdd1501de701e7d5fb6c34da9"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.312100 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xg4dj" event={"ID":"59a2dfdf-8626-45d9-a37b-91272ecbc570","Type":"ContainerStarted","Data":"a8974ea658ab00fda5797f8ea8d0a13fd60c609b11c7e075c75a52b28c443b5d"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.320921 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7119330c-76fa-4850-9372-f3dd55c20247","Type":"ContainerStarted","Data":"59dc8b64f1ca9470f5d928b6274127084fc0d5ca0dcf93642ebd079f90104f4e"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.322996 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd0776fb-8edf-4508-98c3-5267fb06cd56","Type":"ContainerStarted","Data":"ac21f9e4efefede49a19626b62e00bcabb8c0c86e89c5bc0e56efd2180c681de"} Nov 26 06:43:32 crc kubenswrapper[4775]: W1126 06:43:32.323898 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c567945_b393_4d26_8916_cb24e2c38e07.slice/crio-425a3db60b9291ea265b0bd3469961f933e22e53920fc416d451735e7529f82a WatchSource:0}: Error finding container 425a3db60b9291ea265b0bd3469961f933e22e53920fc416d451735e7529f82a: Status 404 returned error can't find the container with id 425a3db60b9291ea265b0bd3469961f933e22e53920fc416d451735e7529f82a Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.328758 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a3529d7-ae55-48f8-9c32-392432471665","Type":"ContainerStarted","Data":"3fb55778f20ebf0d221428210ea408ecbb6bcc6c9cb98329a0cb78993ee05fc4"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.333463 4775 generic.go:334] "Generic (PLEG): container finished" podID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerID="8b2efc0704e9f5003ff0b19894d0861682d44396a6b6b8c21a3b29a2c7fef314" exitCode=0 Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.333527 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" event={"ID":"ef02f1ef-024a-4c85-826f-7b0c4749dbdd","Type":"ContainerDied","Data":"8b2efc0704e9f5003ff0b19894d0861682d44396a6b6b8c21a3b29a2c7fef314"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.333830 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" event={"ID":"ef02f1ef-024a-4c85-826f-7b0c4749dbdd","Type":"ContainerStarted","Data":"512fdbe0f2201de777a9ed7a34e2d5d03f0e05267cce70455224d87aded20077"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.336552 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1418942b-2d58-49c9-b24f-9bb582ea3ed8","Type":"ContainerStarted","Data":"30954ef2d49e869b1da789bea2bcbf118fb6cc4de22e028736e2cdb993bb2e58"} Nov 26 06:43:32 crc kubenswrapper[4775]: I1126 06:43:32.347179 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xg4dj" podStartSLOduration=2.347157923 podStartE2EDuration="2.347157923s" podCreationTimestamp="2025-11-26 06:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:32.334739191 +0000 UTC m=+1095.696043163" watchObservedRunningTime="2025-11-26 06:43:32.347157923 +0000 UTC m=+1095.708461875" Nov 26 06:43:33 crc kubenswrapper[4775]: I1126 06:43:33.357298 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6prq8" event={"ID":"5c567945-b393-4d26-8916-cb24e2c38e07","Type":"ContainerStarted","Data":"48a33ff119297c23aba3d9c9db8e7030b84892f7ee4b25bd6d00be5f8e8829b2"} Nov 26 06:43:33 crc kubenswrapper[4775]: I1126 06:43:33.357601 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6prq8" event={"ID":"5c567945-b393-4d26-8916-cb24e2c38e07","Type":"ContainerStarted","Data":"425a3db60b9291ea265b0bd3469961f933e22e53920fc416d451735e7529f82a"} Nov 26 06:43:33 crc kubenswrapper[4775]: I1126 06:43:33.363525 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" event={"ID":"ef02f1ef-024a-4c85-826f-7b0c4749dbdd","Type":"ContainerStarted","Data":"20a19d0b0eb11331e934143d7cff3572c583c8581e65550c8b8a42cce159863a"} Nov 26 06:43:33 crc kubenswrapper[4775]: I1126 06:43:33.363775 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:33 crc kubenswrapper[4775]: I1126 06:43:33.377211 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6prq8" podStartSLOduration=2.377193603 podStartE2EDuration="2.377193603s" podCreationTimestamp="2025-11-26 06:43:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:33.37469409 +0000 UTC m=+1096.735998042" watchObservedRunningTime="2025-11-26 06:43:33.377193603 +0000 UTC m=+1096.738497555" Nov 26 06:43:34 crc kubenswrapper[4775]: I1126 06:43:34.235830 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" podStartSLOduration=4.235807391 podStartE2EDuration="4.235807391s" podCreationTimestamp="2025-11-26 06:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:33.405399651 +0000 UTC m=+1096.766703603" watchObservedRunningTime="2025-11-26 06:43:34.235807391 +0000 UTC m=+1097.597111333" Nov 26 06:43:34 crc kubenswrapper[4775]: I1126 06:43:34.236452 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:43:34 crc kubenswrapper[4775]: I1126 06:43:34.293653 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.389895 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a3529d7-ae55-48f8-9c32-392432471665","Type":"ContainerStarted","Data":"855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485"} Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.390173 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a3529d7-ae55-48f8-9c32-392432471665","Type":"ContainerStarted","Data":"4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855"} Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.390061 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-metadata" containerID="cri-o://855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485" gracePeriod=30 Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.389969 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-log" containerID="cri-o://4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855" gracePeriod=30 Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.391358 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1418942b-2d58-49c9-b24f-9bb582ea3ed8","Type":"ContainerStarted","Data":"e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a"} Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.403591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7119330c-76fa-4850-9372-f3dd55c20247","Type":"ContainerStarted","Data":"61b19e016fd1ac19b543a0e8b4634d490a88bc1141759d2669269ff25c55e8ed"} Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.403640 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7119330c-76fa-4850-9372-f3dd55c20247","Type":"ContainerStarted","Data":"11d428ed975424b2e7c61b70a3b5a07a36218bd10c10c0bd1bb6973b6540c99a"} Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.407255 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd0776fb-8edf-4508-98c3-5267fb06cd56","Type":"ContainerStarted","Data":"26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b"} Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.407392 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="dd0776fb-8edf-4508-98c3-5267fb06cd56" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b" gracePeriod=30 Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.414649 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4387297119999998 podStartE2EDuration="5.414626948s" podCreationTimestamp="2025-11-26 06:43:30 +0000 UTC" firstStartedPulling="2025-11-26 06:43:31.693342617 +0000 UTC m=+1095.054646569" lastFinishedPulling="2025-11-26 06:43:34.669239853 +0000 UTC m=+1098.030543805" observedRunningTime="2025-11-26 06:43:35.411157111 +0000 UTC m=+1098.772461063" watchObservedRunningTime="2025-11-26 06:43:35.414626948 +0000 UTC m=+1098.775930900" Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.457212 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.391815074 podStartE2EDuration="5.457189746s" podCreationTimestamp="2025-11-26 06:43:30 +0000 UTC" firstStartedPulling="2025-11-26 06:43:31.599726357 +0000 UTC m=+1094.961030309" lastFinishedPulling="2025-11-26 06:43:34.665101029 +0000 UTC m=+1098.026404981" observedRunningTime="2025-11-26 06:43:35.431948063 +0000 UTC m=+1098.793252005" watchObservedRunningTime="2025-11-26 06:43:35.457189746 +0000 UTC m=+1098.818493718" Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.473054 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.650613982 podStartE2EDuration="5.473030354s" podCreationTimestamp="2025-11-26 06:43:30 +0000 UTC" firstStartedPulling="2025-11-26 06:43:31.84319226 +0000 UTC m=+1095.204496212" lastFinishedPulling="2025-11-26 06:43:34.665608632 +0000 UTC m=+1098.026912584" observedRunningTime="2025-11-26 06:43:35.446052707 +0000 UTC m=+1098.807356649" watchObservedRunningTime="2025-11-26 06:43:35.473030354 +0000 UTC m=+1098.834334316" Nov 26 06:43:35 crc kubenswrapper[4775]: I1126 06:43:35.486576 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.79070619 podStartE2EDuration="5.486559334s" podCreationTimestamp="2025-11-26 06:43:30 +0000 UTC" firstStartedPulling="2025-11-26 06:43:31.968898846 +0000 UTC m=+1095.330202798" lastFinishedPulling="2025-11-26 06:43:34.66475199 +0000 UTC m=+1098.026055942" observedRunningTime="2025-11-26 06:43:35.463932476 +0000 UTC m=+1098.825236428" watchObservedRunningTime="2025-11-26 06:43:35.486559334 +0000 UTC m=+1098.847863276" Nov 26 06:43:36 crc kubenswrapper[4775]: I1126 06:43:36.121791 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 06:43:36 crc kubenswrapper[4775]: I1126 06:43:36.122117 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 06:43:36 crc kubenswrapper[4775]: I1126 06:43:36.303981 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 26 06:43:36 crc kubenswrapper[4775]: I1126 06:43:36.336812 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:43:36 crc kubenswrapper[4775]: I1126 06:43:36.417729 4775 generic.go:334] "Generic (PLEG): container finished" podID="6a3529d7-ae55-48f8-9c32-392432471665" containerID="4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855" exitCode=143 Nov 26 06:43:36 crc kubenswrapper[4775]: I1126 06:43:36.417757 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a3529d7-ae55-48f8-9c32-392432471665","Type":"ContainerDied","Data":"4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855"} Nov 26 06:43:39 crc kubenswrapper[4775]: I1126 06:43:39.465933 4775 generic.go:334] "Generic (PLEG): container finished" podID="59a2dfdf-8626-45d9-a37b-91272ecbc570" containerID="5cf4f039200a4d89cb81f8faf7b4ffb773a750ebdd1501de701e7d5fb6c34da9" exitCode=0 Nov 26 06:43:39 crc kubenswrapper[4775]: I1126 06:43:39.466432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xg4dj" event={"ID":"59a2dfdf-8626-45d9-a37b-91272ecbc570","Type":"ContainerDied","Data":"5cf4f039200a4d89cb81f8faf7b4ffb773a750ebdd1501de701e7d5fb6c34da9"} Nov 26 06:43:40 crc kubenswrapper[4775]: I1126 06:43:40.479928 4775 generic.go:334] "Generic (PLEG): container finished" podID="5c567945-b393-4d26-8916-cb24e2c38e07" containerID="48a33ff119297c23aba3d9c9db8e7030b84892f7ee4b25bd6d00be5f8e8829b2" exitCode=0 Nov 26 06:43:40 crc kubenswrapper[4775]: I1126 06:43:40.480045 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6prq8" event={"ID":"5c567945-b393-4d26-8916-cb24e2c38e07","Type":"ContainerDied","Data":"48a33ff119297c23aba3d9c9db8e7030b84892f7ee4b25bd6d00be5f8e8829b2"} Nov 26 06:43:40 crc kubenswrapper[4775]: I1126 06:43:40.944962 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.081160 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.081206 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.096501 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-scripts\") pod \"59a2dfdf-8626-45d9-a37b-91272ecbc570\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.096633 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8njqd\" (UniqueName: \"kubernetes.io/projected/59a2dfdf-8626-45d9-a37b-91272ecbc570-kube-api-access-8njqd\") pod \"59a2dfdf-8626-45d9-a37b-91272ecbc570\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.096984 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-combined-ca-bundle\") pod \"59a2dfdf-8626-45d9-a37b-91272ecbc570\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.097231 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-config-data\") pod \"59a2dfdf-8626-45d9-a37b-91272ecbc570\" (UID: \"59a2dfdf-8626-45d9-a37b-91272ecbc570\") " Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.103415 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59a2dfdf-8626-45d9-a37b-91272ecbc570-kube-api-access-8njqd" (OuterVolumeSpecName: "kube-api-access-8njqd") pod "59a2dfdf-8626-45d9-a37b-91272ecbc570" (UID: "59a2dfdf-8626-45d9-a37b-91272ecbc570"). InnerVolumeSpecName "kube-api-access-8njqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.103444 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-scripts" (OuterVolumeSpecName: "scripts") pod "59a2dfdf-8626-45d9-a37b-91272ecbc570" (UID: "59a2dfdf-8626-45d9-a37b-91272ecbc570"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.127462 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-config-data" (OuterVolumeSpecName: "config-data") pod "59a2dfdf-8626-45d9-a37b-91272ecbc570" (UID: "59a2dfdf-8626-45d9-a37b-91272ecbc570"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.139645 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59a2dfdf-8626-45d9-a37b-91272ecbc570" (UID: "59a2dfdf-8626-45d9-a37b-91272ecbc570"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.199410 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.199456 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8njqd\" (UniqueName: \"kubernetes.io/projected/59a2dfdf-8626-45d9-a37b-91272ecbc570-kube-api-access-8njqd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.199470 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.199482 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59a2dfdf-8626-45d9-a37b-91272ecbc570-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.303253 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.345023 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.357253 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.434536 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-bchpj"] Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.435298 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" podUID="4770209f-010d-41f0-b211-1bab5637242c" containerName="dnsmasq-dns" containerID="cri-o://b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585" gracePeriod=10 Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.514696 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xg4dj" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.515157 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xg4dj" event={"ID":"59a2dfdf-8626-45d9-a37b-91272ecbc570","Type":"ContainerDied","Data":"a8974ea658ab00fda5797f8ea8d0a13fd60c609b11c7e075c75a52b28c443b5d"} Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.515176 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8974ea658ab00fda5797f8ea8d0a13fd60c609b11c7e075c75a52b28c443b5d" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.617560 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.699376 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.699632 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-log" containerID="cri-o://11d428ed975424b2e7c61b70a3b5a07a36218bd10c10c0bd1bb6973b6540c99a" gracePeriod=30 Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.700102 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-api" containerID="cri-o://61b19e016fd1ac19b543a0e8b4634d490a88bc1141759d2669269ff25c55e8ed" gracePeriod=30 Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.704572 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": EOF" Nov 26 06:43:41 crc kubenswrapper[4775]: I1126 06:43:41.706555 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.184:8774/\": EOF" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.113789 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.122018 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.210817 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221466 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5m8b\" (UniqueName: \"kubernetes.io/projected/5c567945-b393-4d26-8916-cb24e2c38e07-kube-api-access-g5m8b\") pod \"5c567945-b393-4d26-8916-cb24e2c38e07\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221516 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-nb\") pod \"4770209f-010d-41f0-b211-1bab5637242c\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221566 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-combined-ca-bundle\") pod \"5c567945-b393-4d26-8916-cb24e2c38e07\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221594 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-swift-storage-0\") pod \"4770209f-010d-41f0-b211-1bab5637242c\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-config\") pod \"4770209f-010d-41f0-b211-1bab5637242c\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221756 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7mgz\" (UniqueName: \"kubernetes.io/projected/4770209f-010d-41f0-b211-1bab5637242c-kube-api-access-r7mgz\") pod \"4770209f-010d-41f0-b211-1bab5637242c\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221808 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-scripts\") pod \"5c567945-b393-4d26-8916-cb24e2c38e07\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221868 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-sb\") pod \"4770209f-010d-41f0-b211-1bab5637242c\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221922 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-config-data\") pod \"5c567945-b393-4d26-8916-cb24e2c38e07\" (UID: \"5c567945-b393-4d26-8916-cb24e2c38e07\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.221949 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-svc\") pod \"4770209f-010d-41f0-b211-1bab5637242c\" (UID: \"4770209f-010d-41f0-b211-1bab5637242c\") " Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.230812 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-scripts" (OuterVolumeSpecName: "scripts") pod "5c567945-b393-4d26-8916-cb24e2c38e07" (UID: "5c567945-b393-4d26-8916-cb24e2c38e07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.231140 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4770209f-010d-41f0-b211-1bab5637242c-kube-api-access-r7mgz" (OuterVolumeSpecName: "kube-api-access-r7mgz") pod "4770209f-010d-41f0-b211-1bab5637242c" (UID: "4770209f-010d-41f0-b211-1bab5637242c"). InnerVolumeSpecName "kube-api-access-r7mgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.255310 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c567945-b393-4d26-8916-cb24e2c38e07-kube-api-access-g5m8b" (OuterVolumeSpecName: "kube-api-access-g5m8b") pod "5c567945-b393-4d26-8916-cb24e2c38e07" (UID: "5c567945-b393-4d26-8916-cb24e2c38e07"). InnerVolumeSpecName "kube-api-access-g5m8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.284376 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c567945-b393-4d26-8916-cb24e2c38e07" (UID: "5c567945-b393-4d26-8916-cb24e2c38e07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.301393 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4770209f-010d-41f0-b211-1bab5637242c" (UID: "4770209f-010d-41f0-b211-1bab5637242c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.314768 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4770209f-010d-41f0-b211-1bab5637242c" (UID: "4770209f-010d-41f0-b211-1bab5637242c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.314940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4770209f-010d-41f0-b211-1bab5637242c" (UID: "4770209f-010d-41f0-b211-1bab5637242c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324636 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324678 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5m8b\" (UniqueName: \"kubernetes.io/projected/5c567945-b393-4d26-8916-cb24e2c38e07-kube-api-access-g5m8b\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324692 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324701 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324709 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324740 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7mgz\" (UniqueName: \"kubernetes.io/projected/4770209f-010d-41f0-b211-1bab5637242c-kube-api-access-r7mgz\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324749 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.324912 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-config-data" (OuterVolumeSpecName: "config-data") pod "5c567945-b393-4d26-8916-cb24e2c38e07" (UID: "5c567945-b393-4d26-8916-cb24e2c38e07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.342349 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-config" (OuterVolumeSpecName: "config") pod "4770209f-010d-41f0-b211-1bab5637242c" (UID: "4770209f-010d-41f0-b211-1bab5637242c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.347242 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4770209f-010d-41f0-b211-1bab5637242c" (UID: "4770209f-010d-41f0-b211-1bab5637242c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.427153 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.427192 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c567945-b393-4d26-8916-cb24e2c38e07-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.427206 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4770209f-010d-41f0-b211-1bab5637242c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.525780 4775 generic.go:334] "Generic (PLEG): container finished" podID="4770209f-010d-41f0-b211-1bab5637242c" containerID="b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585" exitCode=0 Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.525853 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" event={"ID":"4770209f-010d-41f0-b211-1bab5637242c","Type":"ContainerDied","Data":"b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585"} Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.525885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" event={"ID":"4770209f-010d-41f0-b211-1bab5637242c","Type":"ContainerDied","Data":"c88b8386087a6b47851e531ad9fed5c0093ca501b4420811d50cc448aeb5f1ad"} Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.525906 4775 scope.go:117] "RemoveContainer" containerID="b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.526085 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-bchpj" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.543383 4775 generic.go:334] "Generic (PLEG): container finished" podID="7119330c-76fa-4850-9372-f3dd55c20247" containerID="11d428ed975424b2e7c61b70a3b5a07a36218bd10c10c0bd1bb6973b6540c99a" exitCode=143 Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.543815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7119330c-76fa-4850-9372-f3dd55c20247","Type":"ContainerDied","Data":"11d428ed975424b2e7c61b70a3b5a07a36218bd10c10c0bd1bb6973b6540c99a"} Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.568006 4775 scope.go:117] "RemoveContainer" containerID="fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.568547 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6prq8" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.569359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6prq8" event={"ID":"5c567945-b393-4d26-8916-cb24e2c38e07","Type":"ContainerDied","Data":"425a3db60b9291ea265b0bd3469961f933e22e53920fc416d451735e7529f82a"} Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.569405 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="425a3db60b9291ea265b0bd3469961f933e22e53920fc416d451735e7529f82a" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.616525 4775 scope.go:117] "RemoveContainer" containerID="b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585" Nov 26 06:43:42 crc kubenswrapper[4775]: E1126 06:43:42.619740 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585\": container with ID starting with b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585 not found: ID does not exist" containerID="b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.619792 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585"} err="failed to get container status \"b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585\": rpc error: code = NotFound desc = could not find container \"b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585\": container with ID starting with b08d3cd89581ecba3a37ff446870de45c93456f40111bb03ed18a1da379a2585 not found: ID does not exist" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.619822 4775 scope.go:117] "RemoveContainer" containerID="fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0" Nov 26 06:43:42 crc kubenswrapper[4775]: E1126 06:43:42.621199 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0\": container with ID starting with fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0 not found: ID does not exist" containerID="fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.621229 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0"} err="failed to get container status \"fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0\": rpc error: code = NotFound desc = could not find container \"fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0\": container with ID starting with fe55a7f235f386a51fec5430f2780609a99ca5087a7287d4af5aa44492e6c4b0 not found: ID does not exist" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.626782 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-bchpj"] Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.646780 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-bchpj"] Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.655788 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 26 06:43:42 crc kubenswrapper[4775]: E1126 06:43:42.656140 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c567945-b393-4d26-8916-cb24e2c38e07" containerName="nova-cell1-conductor-db-sync" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656158 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c567945-b393-4d26-8916-cb24e2c38e07" containerName="nova-cell1-conductor-db-sync" Nov 26 06:43:42 crc kubenswrapper[4775]: E1126 06:43:42.656168 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4770209f-010d-41f0-b211-1bab5637242c" containerName="init" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656175 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4770209f-010d-41f0-b211-1bab5637242c" containerName="init" Nov 26 06:43:42 crc kubenswrapper[4775]: E1126 06:43:42.656197 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a2dfdf-8626-45d9-a37b-91272ecbc570" containerName="nova-manage" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656203 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a2dfdf-8626-45d9-a37b-91272ecbc570" containerName="nova-manage" Nov 26 06:43:42 crc kubenswrapper[4775]: E1126 06:43:42.656224 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4770209f-010d-41f0-b211-1bab5637242c" containerName="dnsmasq-dns" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656230 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4770209f-010d-41f0-b211-1bab5637242c" containerName="dnsmasq-dns" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656386 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c567945-b393-4d26-8916-cb24e2c38e07" containerName="nova-cell1-conductor-db-sync" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656399 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4770209f-010d-41f0-b211-1bab5637242c" containerName="dnsmasq-dns" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.656417 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="59a2dfdf-8626-45d9-a37b-91272ecbc570" containerName="nova-manage" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.657111 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.659870 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.665989 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.737910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64m4g\" (UniqueName: \"kubernetes.io/projected/ec4083e2-96ad-420c-a66f-0cb0f274867a-kube-api-access-64m4g\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.737967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4083e2-96ad-420c-a66f-0cb0f274867a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.738080 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4083e2-96ad-420c-a66f-0cb0f274867a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.840314 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4083e2-96ad-420c-a66f-0cb0f274867a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.840449 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64m4g\" (UniqueName: \"kubernetes.io/projected/ec4083e2-96ad-420c-a66f-0cb0f274867a-kube-api-access-64m4g\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.840471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4083e2-96ad-420c-a66f-0cb0f274867a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.844945 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec4083e2-96ad-420c-a66f-0cb0f274867a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.847424 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec4083e2-96ad-420c-a66f-0cb0f274867a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.857602 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64m4g\" (UniqueName: \"kubernetes.io/projected/ec4083e2-96ad-420c-a66f-0cb0f274867a-kube-api-access-64m4g\") pod \"nova-cell1-conductor-0\" (UID: \"ec4083e2-96ad-420c-a66f-0cb0f274867a\") " pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:42 crc kubenswrapper[4775]: I1126 06:43:42.985204 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:43 crc kubenswrapper[4775]: I1126 06:43:43.341254 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4770209f-010d-41f0-b211-1bab5637242c" path="/var/lib/kubelet/pods/4770209f-010d-41f0-b211-1bab5637242c/volumes" Nov 26 06:43:43 crc kubenswrapper[4775]: I1126 06:43:43.580534 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" containerName="nova-scheduler-scheduler" containerID="cri-o://e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a" gracePeriod=30 Nov 26 06:43:43 crc kubenswrapper[4775]: W1126 06:43:43.601912 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec4083e2_96ad_420c_a66f_0cb0f274867a.slice/crio-f58f0a16c26933fb32b4c63eac71e01a612231b6f8ac5dc03329930189959573 WatchSource:0}: Error finding container f58f0a16c26933fb32b4c63eac71e01a612231b6f8ac5dc03329930189959573: Status 404 returned error can't find the container with id f58f0a16c26933fb32b4c63eac71e01a612231b6f8ac5dc03329930189959573 Nov 26 06:43:43 crc kubenswrapper[4775]: I1126 06:43:43.603149 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 26 06:43:44 crc kubenswrapper[4775]: I1126 06:43:44.591102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ec4083e2-96ad-420c-a66f-0cb0f274867a","Type":"ContainerStarted","Data":"2fd97af3e3b7de654773b7d2bca8d710c4c399121635e2b660b97dc28e188ecb"} Nov 26 06:43:44 crc kubenswrapper[4775]: I1126 06:43:44.591419 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:44 crc kubenswrapper[4775]: I1126 06:43:44.591430 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ec4083e2-96ad-420c-a66f-0cb0f274867a","Type":"ContainerStarted","Data":"f58f0a16c26933fb32b4c63eac71e01a612231b6f8ac5dc03329930189959573"} Nov 26 06:43:44 crc kubenswrapper[4775]: I1126 06:43:44.611826 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.611801609 podStartE2EDuration="2.611801609s" podCreationTimestamp="2025-11-26 06:43:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:44.60535747 +0000 UTC m=+1107.966661432" watchObservedRunningTime="2025-11-26 06:43:44.611801609 +0000 UTC m=+1107.973105581" Nov 26 06:43:46 crc kubenswrapper[4775]: E1126 06:43:46.305092 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 06:43:46 crc kubenswrapper[4775]: E1126 06:43:46.307295 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 06:43:46 crc kubenswrapper[4775]: E1126 06:43:46.308990 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 26 06:43:46 crc kubenswrapper[4775]: E1126 06:43:46.309178 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" containerName="nova-scheduler-scheduler" Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.516462 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.622177 4775 generic.go:334] "Generic (PLEG): container finished" podID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" containerID="e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a" exitCode=0 Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.622218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1418942b-2d58-49c9-b24f-9bb582ea3ed8","Type":"ContainerDied","Data":"e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a"} Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.797784 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.913996 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4q4n\" (UniqueName: \"kubernetes.io/projected/1418942b-2d58-49c9-b24f-9bb582ea3ed8-kube-api-access-v4q4n\") pod \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.914090 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-config-data\") pod \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.914124 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-combined-ca-bundle\") pod \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\" (UID: \"1418942b-2d58-49c9-b24f-9bb582ea3ed8\") " Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.920412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1418942b-2d58-49c9-b24f-9bb582ea3ed8-kube-api-access-v4q4n" (OuterVolumeSpecName: "kube-api-access-v4q4n") pod "1418942b-2d58-49c9-b24f-9bb582ea3ed8" (UID: "1418942b-2d58-49c9-b24f-9bb582ea3ed8"). InnerVolumeSpecName "kube-api-access-v4q4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.944372 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1418942b-2d58-49c9-b24f-9bb582ea3ed8" (UID: "1418942b-2d58-49c9-b24f-9bb582ea3ed8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:46 crc kubenswrapper[4775]: I1126 06:43:46.944853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-config-data" (OuterVolumeSpecName: "config-data") pod "1418942b-2d58-49c9-b24f-9bb582ea3ed8" (UID: "1418942b-2d58-49c9-b24f-9bb582ea3ed8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.016256 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4q4n\" (UniqueName: \"kubernetes.io/projected/1418942b-2d58-49c9-b24f-9bb582ea3ed8-kube-api-access-v4q4n\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.016294 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.016309 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1418942b-2d58-49c9-b24f-9bb582ea3ed8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.633678 4775 generic.go:334] "Generic (PLEG): container finished" podID="7119330c-76fa-4850-9372-f3dd55c20247" containerID="61b19e016fd1ac19b543a0e8b4634d490a88bc1141759d2669269ff25c55e8ed" exitCode=0 Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.634026 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7119330c-76fa-4850-9372-f3dd55c20247","Type":"ContainerDied","Data":"61b19e016fd1ac19b543a0e8b4634d490a88bc1141759d2669269ff25c55e8ed"} Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.635672 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1418942b-2d58-49c9-b24f-9bb582ea3ed8","Type":"ContainerDied","Data":"30954ef2d49e869b1da789bea2bcbf118fb6cc4de22e028736e2cdb993bb2e58"} Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.635731 4775 scope.go:117] "RemoveContainer" containerID="e69b2cebe191aa629f8039961bf0a48455387197e64987b624ad82ba3415425a" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.635861 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.676127 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.693616 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.705348 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:47 crc kubenswrapper[4775]: E1126 06:43:47.705658 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" containerName="nova-scheduler-scheduler" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.705669 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" containerName="nova-scheduler-scheduler" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.705939 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" containerName="nova-scheduler-scheduler" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.706491 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.707959 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.716628 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.803432 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.836803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9s74\" (UniqueName: \"kubernetes.io/projected/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-kube-api-access-z9s74\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.837169 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.837217 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-config-data\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.963323 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7119330c-76fa-4850-9372-f3dd55c20247-logs\") pod \"7119330c-76fa-4850-9372-f3dd55c20247\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.963428 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jm94\" (UniqueName: \"kubernetes.io/projected/7119330c-76fa-4850-9372-f3dd55c20247-kube-api-access-9jm94\") pod \"7119330c-76fa-4850-9372-f3dd55c20247\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.963503 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-combined-ca-bundle\") pod \"7119330c-76fa-4850-9372-f3dd55c20247\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.963582 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-config-data\") pod \"7119330c-76fa-4850-9372-f3dd55c20247\" (UID: \"7119330c-76fa-4850-9372-f3dd55c20247\") " Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.964126 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7119330c-76fa-4850-9372-f3dd55c20247-logs" (OuterVolumeSpecName: "logs") pod "7119330c-76fa-4850-9372-f3dd55c20247" (UID: "7119330c-76fa-4850-9372-f3dd55c20247"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.964430 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.964559 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-config-data\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.964745 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9s74\" (UniqueName: \"kubernetes.io/projected/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-kube-api-access-z9s74\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.964936 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7119330c-76fa-4850-9372-f3dd55c20247-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.975083 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7119330c-76fa-4850-9372-f3dd55c20247-kube-api-access-9jm94" (OuterVolumeSpecName: "kube-api-access-9jm94") pod "7119330c-76fa-4850-9372-f3dd55c20247" (UID: "7119330c-76fa-4850-9372-f3dd55c20247"). InnerVolumeSpecName "kube-api-access-9jm94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.977781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-config-data\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.987620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.991108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9s74\" (UniqueName: \"kubernetes.io/projected/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-kube-api-access-z9s74\") pod \"nova-scheduler-0\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " pod="openstack/nova-scheduler-0" Nov 26 06:43:47 crc kubenswrapper[4775]: I1126 06:43:47.993288 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7119330c-76fa-4850-9372-f3dd55c20247" (UID: "7119330c-76fa-4850-9372-f3dd55c20247"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.008187 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-config-data" (OuterVolumeSpecName: "config-data") pod "7119330c-76fa-4850-9372-f3dd55c20247" (UID: "7119330c-76fa-4850-9372-f3dd55c20247"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.066557 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jm94\" (UniqueName: \"kubernetes.io/projected/7119330c-76fa-4850-9372-f3dd55c20247-kube-api-access-9jm94\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.066599 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.066611 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7119330c-76fa-4850-9372-f3dd55c20247-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.116964 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.652998 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7119330c-76fa-4850-9372-f3dd55c20247","Type":"ContainerDied","Data":"59dc8b64f1ca9470f5d928b6274127084fc0d5ca0dcf93642ebd079f90104f4e"} Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.653056 4775 scope.go:117] "RemoveContainer" containerID="61b19e016fd1ac19b543a0e8b4634d490a88bc1141759d2669269ff25c55e8ed" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.653239 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.656042 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.750818 4775 scope.go:117] "RemoveContainer" containerID="11d428ed975424b2e7c61b70a3b5a07a36218bd10c10c0bd1bb6973b6540c99a" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.756812 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.769692 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.782288 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:48 crc kubenswrapper[4775]: E1126 06:43:48.782678 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-log" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.782736 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-log" Nov 26 06:43:48 crc kubenswrapper[4775]: E1126 06:43:48.782771 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-api" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.782777 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-api" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.782945 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-log" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.782981 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7119330c-76fa-4850-9372-f3dd55c20247" containerName="nova-api-api" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.784984 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.787231 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.792043 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.884113 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.884200 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60149bf-8618-4ecd-9cd7-80d59881a7b6-logs\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.884244 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-config-data\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.884295 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44fvx\" (UniqueName: \"kubernetes.io/projected/c60149bf-8618-4ecd-9cd7-80d59881a7b6-kube-api-access-44fvx\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.986540 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44fvx\" (UniqueName: \"kubernetes.io/projected/c60149bf-8618-4ecd-9cd7-80d59881a7b6-kube-api-access-44fvx\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.986658 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.986744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60149bf-8618-4ecd-9cd7-80d59881a7b6-logs\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.986781 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-config-data\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.987293 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60149bf-8618-4ecd-9cd7-80d59881a7b6-logs\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:48 crc kubenswrapper[4775]: I1126 06:43:48.991955 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-config-data\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.003638 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.008265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44fvx\" (UniqueName: \"kubernetes.io/projected/c60149bf-8618-4ecd-9cd7-80d59881a7b6-kube-api-access-44fvx\") pod \"nova-api-0\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " pod="openstack/nova-api-0" Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.154357 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.350435 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1418942b-2d58-49c9-b24f-9bb582ea3ed8" path="/var/lib/kubelet/pods/1418942b-2d58-49c9-b24f-9bb582ea3ed8/volumes" Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.351190 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7119330c-76fa-4850-9372-f3dd55c20247" path="/var/lib/kubelet/pods/7119330c-76fa-4850-9372-f3dd55c20247/volumes" Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.663038 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.667965 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00","Type":"ContainerStarted","Data":"57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae"} Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.668011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00","Type":"ContainerStarted","Data":"91f2ddb7b2d3843a1b92f31471723eaa1a387659388591da37acec746e32e853"} Nov 26 06:43:49 crc kubenswrapper[4775]: I1126 06:43:49.693103 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.69308545 podStartE2EDuration="2.69308545s" podCreationTimestamp="2025-11-26 06:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:49.681012781 +0000 UTC m=+1113.042316743" watchObservedRunningTime="2025-11-26 06:43:49.69308545 +0000 UTC m=+1113.054389402" Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.535074 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.535568 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="633ccd9d-4b8e-4797-89f9-7412c416b35b" containerName="kube-state-metrics" containerID="cri-o://f9ef220083717002d290f22b3a464b503d762b539cfc24ac41155ce7b03aaef3" gracePeriod=30 Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.733913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c60149bf-8618-4ecd-9cd7-80d59881a7b6","Type":"ContainerStarted","Data":"e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b"} Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.733964 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c60149bf-8618-4ecd-9cd7-80d59881a7b6","Type":"ContainerStarted","Data":"cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299"} Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.733981 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c60149bf-8618-4ecd-9cd7-80d59881a7b6","Type":"ContainerStarted","Data":"9b653d5114454fb6073a515b6bdf6a3fd6dadbede69feb91b1fff970ea02c585"} Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.754347 4775 generic.go:334] "Generic (PLEG): container finished" podID="633ccd9d-4b8e-4797-89f9-7412c416b35b" containerID="f9ef220083717002d290f22b3a464b503d762b539cfc24ac41155ce7b03aaef3" exitCode=2 Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.756330 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"633ccd9d-4b8e-4797-89f9-7412c416b35b","Type":"ContainerDied","Data":"f9ef220083717002d290f22b3a464b503d762b539cfc24ac41155ce7b03aaef3"} Nov 26 06:43:50 crc kubenswrapper[4775]: I1126 06:43:50.788693 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.788673204 podStartE2EDuration="2.788673204s" podCreationTimestamp="2025-11-26 06:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:43:50.773675737 +0000 UTC m=+1114.134979689" watchObservedRunningTime="2025-11-26 06:43:50.788673204 +0000 UTC m=+1114.149977156" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.172686 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.234825 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gnrp\" (UniqueName: \"kubernetes.io/projected/633ccd9d-4b8e-4797-89f9-7412c416b35b-kube-api-access-4gnrp\") pod \"633ccd9d-4b8e-4797-89f9-7412c416b35b\" (UID: \"633ccd9d-4b8e-4797-89f9-7412c416b35b\") " Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.255156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/633ccd9d-4b8e-4797-89f9-7412c416b35b-kube-api-access-4gnrp" (OuterVolumeSpecName: "kube-api-access-4gnrp") pod "633ccd9d-4b8e-4797-89f9-7412c416b35b" (UID: "633ccd9d-4b8e-4797-89f9-7412c416b35b"). InnerVolumeSpecName "kube-api-access-4gnrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.336515 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gnrp\" (UniqueName: \"kubernetes.io/projected/633ccd9d-4b8e-4797-89f9-7412c416b35b-kube-api-access-4gnrp\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.420333 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.420402 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.420450 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.421246 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e3a269c3e7b57838eb3de922a5f67cb80c1084327864aee098382792cdcca9f"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.421315 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://8e3a269c3e7b57838eb3de922a5f67cb80c1084327864aee098382792cdcca9f" gracePeriod=600 Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.764689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"633ccd9d-4b8e-4797-89f9-7412c416b35b","Type":"ContainerDied","Data":"71f48b6f463433d07630a6a6f21c8af9a1f079643e398ec289111cf0d362ec72"} Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.764724 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.765047 4775 scope.go:117] "RemoveContainer" containerID="f9ef220083717002d290f22b3a464b503d762b539cfc24ac41155ce7b03aaef3" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.770800 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="8e3a269c3e7b57838eb3de922a5f67cb80c1084327864aee098382792cdcca9f" exitCode=0 Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.770868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"8e3a269c3e7b57838eb3de922a5f67cb80c1084327864aee098382792cdcca9f"} Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.770913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"8edde92e1ddfc192e6ae3206422c2d764bb7e89d2f7777b6c406add7b5a50f48"} Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.787748 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.798779 4775 scope.go:117] "RemoveContainer" containerID="c257f11ab63c2232fe4ea03480103fc149650fe9c5d870044eeec21f9ccaaf75" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.803627 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.815377 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:43:51 crc kubenswrapper[4775]: E1126 06:43:51.815897 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633ccd9d-4b8e-4797-89f9-7412c416b35b" containerName="kube-state-metrics" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.815924 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="633ccd9d-4b8e-4797-89f9-7412c416b35b" containerName="kube-state-metrics" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.816213 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="633ccd9d-4b8e-4797-89f9-7412c416b35b" containerName="kube-state-metrics" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.817045 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.819340 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.819889 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.844785 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.948998 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.949062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.949229 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:51 crc kubenswrapper[4775]: I1126 06:43:51.949256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4226\" (UniqueName: \"kubernetes.io/projected/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-api-access-d4226\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.050665 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.050746 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.050929 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.050954 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4226\" (UniqueName: \"kubernetes.io/projected/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-api-access-d4226\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.056233 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.057005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.057188 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2427f2bd-7928-4ff9-bc65-23bc48225184-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.072534 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4226\" (UniqueName: \"kubernetes.io/projected/2427f2bd-7928-4ff9-bc65-23bc48225184-kube-api-access-d4226\") pod \"kube-state-metrics-0\" (UID: \"2427f2bd-7928-4ff9-bc65-23bc48225184\") " pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.136994 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.585967 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.586536 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-central-agent" containerID="cri-o://d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811" gracePeriod=30 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.586567 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="proxy-httpd" containerID="cri-o://8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf" gracePeriod=30 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.586593 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="sg-core" containerID="cri-o://4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c" gracePeriod=30 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.586662 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-notification-agent" containerID="cri-o://fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7" gracePeriod=30 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.655752 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 26 06:43:52 crc kubenswrapper[4775]: W1126 06:43:52.660785 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2427f2bd_7928_4ff9_bc65_23bc48225184.slice/crio-647c2b8daf1f5ae670a4a6da910f9f9c9d50f2902a2c7d08039d352ce56ca647 WatchSource:0}: Error finding container 647c2b8daf1f5ae670a4a6da910f9f9c9d50f2902a2c7d08039d352ce56ca647: Status 404 returned error can't find the container with id 647c2b8daf1f5ae670a4a6da910f9f9c9d50f2902a2c7d08039d352ce56ca647 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.784766 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2427f2bd-7928-4ff9-bc65-23bc48225184","Type":"ContainerStarted","Data":"647c2b8daf1f5ae670a4a6da910f9f9c9d50f2902a2c7d08039d352ce56ca647"} Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.788035 4775 generic.go:334] "Generic (PLEG): container finished" podID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerID="8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf" exitCode=0 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.788340 4775 generic.go:334] "Generic (PLEG): container finished" podID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerID="4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c" exitCode=2 Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.788095 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerDied","Data":"8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf"} Nov 26 06:43:52 crc kubenswrapper[4775]: I1126 06:43:52.788414 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerDied","Data":"4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c"} Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.016959 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.118599 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.338682 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="633ccd9d-4b8e-4797-89f9-7412c416b35b" path="/var/lib/kubelet/pods/633ccd9d-4b8e-4797-89f9-7412c416b35b/volumes" Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.801507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2427f2bd-7928-4ff9-bc65-23bc48225184","Type":"ContainerStarted","Data":"83fa83735ba05f236c2da9dfca01f25227a89f51df253c0097bf497b12e39c9a"} Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.801879 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.804025 4775 generic.go:334] "Generic (PLEG): container finished" podID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerID="d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811" exitCode=0 Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.804152 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerDied","Data":"d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811"} Nov 26 06:43:53 crc kubenswrapper[4775]: I1126 06:43:53.833807 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.443223359 podStartE2EDuration="2.833782699s" podCreationTimestamp="2025-11-26 06:43:51 +0000 UTC" firstStartedPulling="2025-11-26 06:43:52.663141575 +0000 UTC m=+1116.024445527" lastFinishedPulling="2025-11-26 06:43:53.053700915 +0000 UTC m=+1116.415004867" observedRunningTime="2025-11-26 06:43:53.828908196 +0000 UTC m=+1117.190212148" watchObservedRunningTime="2025-11-26 06:43:53.833782699 +0000 UTC m=+1117.195086671" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.526679 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596459 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94xzh\" (UniqueName: \"kubernetes.io/projected/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-kube-api-access-94xzh\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596511 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-combined-ca-bundle\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596608 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-scripts\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596687 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-config-data\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596736 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-log-httpd\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596780 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-sg-core-conf-yaml\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.596799 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-run-httpd\") pod \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\" (UID: \"c972fd9f-3d56-460c-b4e8-fac121e0c1ee\") " Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.597292 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.598283 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.602264 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.602312 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.603310 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-scripts" (OuterVolumeSpecName: "scripts") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.618882 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-kube-api-access-94xzh" (OuterVolumeSpecName: "kube-api-access-94xzh") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "kube-api-access-94xzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.634474 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.696303 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.704583 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.704618 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.704631 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94xzh\" (UniqueName: \"kubernetes.io/projected/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-kube-api-access-94xzh\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.704642 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.707965 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-config-data" (OuterVolumeSpecName: "config-data") pod "c972fd9f-3d56-460c-b4e8-fac121e0c1ee" (UID: "c972fd9f-3d56-460c-b4e8-fac121e0c1ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.805968 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c972fd9f-3d56-460c-b4e8-fac121e0c1ee-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.814845 4775 generic.go:334] "Generic (PLEG): container finished" podID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerID="fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7" exitCode=0 Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.814903 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.814948 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerDied","Data":"fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7"} Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.814979 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c972fd9f-3d56-460c-b4e8-fac121e0c1ee","Type":"ContainerDied","Data":"e2253269920d3c5f264f18afc2f71d78ad7e6f83da65839dc69e741750b90c1a"} Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.814997 4775 scope.go:117] "RemoveContainer" containerID="8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.834681 4775 scope.go:117] "RemoveContainer" containerID="4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.860124 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.872075 4775 scope.go:117] "RemoveContainer" containerID="fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.877239 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.889764 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.890235 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="sg-core" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890256 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="sg-core" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.890273 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-central-agent" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890280 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-central-agent" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.890303 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="proxy-httpd" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890309 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="proxy-httpd" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.890318 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-notification-agent" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890334 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-notification-agent" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890528 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="proxy-httpd" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890551 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="sg-core" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890566 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-notification-agent" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.890581 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" containerName="ceilometer-central-agent" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.892222 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.897840 4775 scope.go:117] "RemoveContainer" containerID="d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.898555 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.906541 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.906778 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.906959 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.933535 4775 scope.go:117] "RemoveContainer" containerID="8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.935260 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf\": container with ID starting with 8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf not found: ID does not exist" containerID="8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.935309 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf"} err="failed to get container status \"8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf\": rpc error: code = NotFound desc = could not find container \"8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf\": container with ID starting with 8060085ced5b7b668113aba6414f196d5451df192b20d41bd4b9b790205179bf not found: ID does not exist" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.935332 4775 scope.go:117] "RemoveContainer" containerID="4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.935631 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c\": container with ID starting with 4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c not found: ID does not exist" containerID="4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.935680 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c"} err="failed to get container status \"4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c\": rpc error: code = NotFound desc = could not find container \"4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c\": container with ID starting with 4f039d89b640187f5638455e9ec7a7c49593cf4d4ba269dad0006dbfd583902c not found: ID does not exist" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.935731 4775 scope.go:117] "RemoveContainer" containerID="fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.936845 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7\": container with ID starting with fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7 not found: ID does not exist" containerID="fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.936874 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7"} err="failed to get container status \"fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7\": rpc error: code = NotFound desc = could not find container \"fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7\": container with ID starting with fd1659368c1c386ea638454aa5119aa64685d75278281ee7637772ccbd5e40d7 not found: ID does not exist" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.936893 4775 scope.go:117] "RemoveContainer" containerID="d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811" Nov 26 06:43:54 crc kubenswrapper[4775]: E1126 06:43:54.937296 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811\": container with ID starting with d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811 not found: ID does not exist" containerID="d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811" Nov 26 06:43:54 crc kubenswrapper[4775]: I1126 06:43:54.937333 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811"} err="failed to get container status \"d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811\": rpc error: code = NotFound desc = could not find container \"d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811\": container with ID starting with d79dc2b94695dd72159bb48b1bc5424a00f34b462f6dcdcf4d4773f660784811 not found: ID does not exist" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009035 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2knrs\" (UniqueName: \"kubernetes.io/projected/74c15881-a4e4-490e-81dc-d6acfae420ab-kube-api-access-2knrs\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009243 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009426 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009536 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-log-httpd\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-scripts\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009613 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-run-httpd\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009638 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-config-data\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.009674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.110952 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-scripts\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.110994 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-run-httpd\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111019 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-config-data\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111048 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111097 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2knrs\" (UniqueName: \"kubernetes.io/projected/74c15881-a4e4-490e-81dc-d6acfae420ab-kube-api-access-2knrs\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111235 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111288 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-log-httpd\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.111849 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-log-httpd\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.112003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-run-httpd\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.115118 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.115401 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-scripts\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.115467 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.117609 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-config-data\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.124044 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.137192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2knrs\" (UniqueName: \"kubernetes.io/projected/74c15881-a4e4-490e-81dc-d6acfae420ab-kube-api-access-2knrs\") pod \"ceilometer-0\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.220887 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.339562 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c972fd9f-3d56-460c-b4e8-fac121e0c1ee" path="/var/lib/kubelet/pods/c972fd9f-3d56-460c-b4e8-fac121e0c1ee/volumes" Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.662849 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:43:55 crc kubenswrapper[4775]: W1126 06:43:55.672163 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74c15881_a4e4_490e_81dc_d6acfae420ab.slice/crio-35d6f056953144549723f9407089092caefce6db649775db7c3f713b63630ff5 WatchSource:0}: Error finding container 35d6f056953144549723f9407089092caefce6db649775db7c3f713b63630ff5: Status 404 returned error can't find the container with id 35d6f056953144549723f9407089092caefce6db649775db7c3f713b63630ff5 Nov 26 06:43:55 crc kubenswrapper[4775]: I1126 06:43:55.831148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerStarted","Data":"35d6f056953144549723f9407089092caefce6db649775db7c3f713b63630ff5"} Nov 26 06:43:56 crc kubenswrapper[4775]: I1126 06:43:56.848533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerStarted","Data":"e45e03bd4bd0416b56133dafeb79943b4471f761aee19dbe23c9e15ed99c3c7b"} Nov 26 06:43:57 crc kubenswrapper[4775]: I1126 06:43:57.860818 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerStarted","Data":"7f1d8222eb05820df1c0e0a107afb2e4a812df9650c61133fdc33692b3ee3ef6"} Nov 26 06:43:57 crc kubenswrapper[4775]: I1126 06:43:57.862076 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerStarted","Data":"f82c48492922e8418e0c4d123c12575b533627198a2f55eb7cddc2f2848eda3a"} Nov 26 06:43:58 crc kubenswrapper[4775]: I1126 06:43:58.118117 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 26 06:43:58 crc kubenswrapper[4775]: I1126 06:43:58.149476 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 26 06:43:58 crc kubenswrapper[4775]: I1126 06:43:58.908468 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 26 06:43:59 crc kubenswrapper[4775]: I1126 06:43:59.156195 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 06:43:59 crc kubenswrapper[4775]: I1126 06:43:59.156424 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 06:43:59 crc kubenswrapper[4775]: I1126 06:43:59.883020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerStarted","Data":"ac8afae7aaf33e61197deda25d909984167c01878f67988ebda3ffc01644b29f"} Nov 26 06:43:59 crc kubenswrapper[4775]: I1126 06:43:59.883095 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:43:59 crc kubenswrapper[4775]: I1126 06:43:59.918690 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.571388958 podStartE2EDuration="5.918671803s" podCreationTimestamp="2025-11-26 06:43:54 +0000 UTC" firstStartedPulling="2025-11-26 06:43:55.674535068 +0000 UTC m=+1119.035839020" lastFinishedPulling="2025-11-26 06:43:59.021817903 +0000 UTC m=+1122.383121865" observedRunningTime="2025-11-26 06:43:59.904455114 +0000 UTC m=+1123.265759066" watchObservedRunningTime="2025-11-26 06:43:59.918671803 +0000 UTC m=+1123.279975755" Nov 26 06:44:00 crc kubenswrapper[4775]: I1126 06:44:00.237911 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:00 crc kubenswrapper[4775]: I1126 06:44:00.237905 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:02 crc kubenswrapper[4775]: I1126 06:44:02.153897 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.882168 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.887380 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.931548 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-combined-ca-bundle\") pod \"6a3529d7-ae55-48f8-9c32-392432471665\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.931913 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs4pb\" (UniqueName: \"kubernetes.io/projected/6a3529d7-ae55-48f8-9c32-392432471665-kube-api-access-vs4pb\") pod \"6a3529d7-ae55-48f8-9c32-392432471665\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.931961 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3529d7-ae55-48f8-9c32-392432471665-logs\") pod \"6a3529d7-ae55-48f8-9c32-392432471665\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.931998 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-combined-ca-bundle\") pod \"dd0776fb-8edf-4508-98c3-5267fb06cd56\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.932035 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-config-data\") pod \"6a3529d7-ae55-48f8-9c32-392432471665\" (UID: \"6a3529d7-ae55-48f8-9c32-392432471665\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.932119 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nf42\" (UniqueName: \"kubernetes.io/projected/dd0776fb-8edf-4508-98c3-5267fb06cd56-kube-api-access-7nf42\") pod \"dd0776fb-8edf-4508-98c3-5267fb06cd56\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.932193 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-config-data\") pod \"dd0776fb-8edf-4508-98c3-5267fb06cd56\" (UID: \"dd0776fb-8edf-4508-98c3-5267fb06cd56\") " Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.935892 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a3529d7-ae55-48f8-9c32-392432471665-logs" (OuterVolumeSpecName: "logs") pod "6a3529d7-ae55-48f8-9c32-392432471665" (UID: "6a3529d7-ae55-48f8-9c32-392432471665"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.939769 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd0776fb-8edf-4508-98c3-5267fb06cd56-kube-api-access-7nf42" (OuterVolumeSpecName: "kube-api-access-7nf42") pod "dd0776fb-8edf-4508-98c3-5267fb06cd56" (UID: "dd0776fb-8edf-4508-98c3-5267fb06cd56"). InnerVolumeSpecName "kube-api-access-7nf42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.945808 4775 generic.go:334] "Generic (PLEG): container finished" podID="dd0776fb-8edf-4508-98c3-5267fb06cd56" containerID="26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b" exitCode=137 Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.945895 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.945884 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd0776fb-8edf-4508-98c3-5267fb06cd56","Type":"ContainerDied","Data":"26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b"} Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.945948 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dd0776fb-8edf-4508-98c3-5267fb06cd56","Type":"ContainerDied","Data":"ac21f9e4efefede49a19626b62e00bcabb8c0c86e89c5bc0e56efd2180c681de"} Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.945978 4775 scope.go:117] "RemoveContainer" containerID="26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.949521 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3529d7-ae55-48f8-9c32-392432471665-kube-api-access-vs4pb" (OuterVolumeSpecName: "kube-api-access-vs4pb") pod "6a3529d7-ae55-48f8-9c32-392432471665" (UID: "6a3529d7-ae55-48f8-9c32-392432471665"). InnerVolumeSpecName "kube-api-access-vs4pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.953002 4775 generic.go:334] "Generic (PLEG): container finished" podID="6a3529d7-ae55-48f8-9c32-392432471665" containerID="855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485" exitCode=137 Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.953057 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a3529d7-ae55-48f8-9c32-392432471665","Type":"ContainerDied","Data":"855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485"} Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.953094 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a3529d7-ae55-48f8-9c32-392432471665","Type":"ContainerDied","Data":"3fb55778f20ebf0d221428210ea408ecbb6bcc6c9cb98329a0cb78993ee05fc4"} Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.953176 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.968826 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-config-data" (OuterVolumeSpecName: "config-data") pod "6a3529d7-ae55-48f8-9c32-392432471665" (UID: "6a3529d7-ae55-48f8-9c32-392432471665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.975112 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a3529d7-ae55-48f8-9c32-392432471665" (UID: "6a3529d7-ae55-48f8-9c32-392432471665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.976737 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd0776fb-8edf-4508-98c3-5267fb06cd56" (UID: "dd0776fb-8edf-4508-98c3-5267fb06cd56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.979043 4775 scope.go:117] "RemoveContainer" containerID="26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b" Nov 26 06:44:05 crc kubenswrapper[4775]: E1126 06:44:05.979586 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b\": container with ID starting with 26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b not found: ID does not exist" containerID="26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.979638 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b"} err="failed to get container status \"26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b\": rpc error: code = NotFound desc = could not find container \"26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b\": container with ID starting with 26f9702e40aa9fe7bead758400e11f630e07aaea9be89913c23388214a8c471b not found: ID does not exist" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.979675 4775 scope.go:117] "RemoveContainer" containerID="855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.981755 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-config-data" (OuterVolumeSpecName: "config-data") pod "dd0776fb-8edf-4508-98c3-5267fb06cd56" (UID: "dd0776fb-8edf-4508-98c3-5267fb06cd56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:05 crc kubenswrapper[4775]: I1126 06:44:05.997641 4775 scope.go:117] "RemoveContainer" containerID="4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.020136 4775 scope.go:117] "RemoveContainer" containerID="855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485" Nov 26 06:44:06 crc kubenswrapper[4775]: E1126 06:44:06.020531 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485\": container with ID starting with 855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485 not found: ID does not exist" containerID="855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.020577 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485"} err="failed to get container status \"855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485\": rpc error: code = NotFound desc = could not find container \"855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485\": container with ID starting with 855a88301051551ef1170f016d7c0afa34fbd69d9626fa4110e1fdc8fb6ce485 not found: ID does not exist" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.020609 4775 scope.go:117] "RemoveContainer" containerID="4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855" Nov 26 06:44:06 crc kubenswrapper[4775]: E1126 06:44:06.021195 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855\": container with ID starting with 4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855 not found: ID does not exist" containerID="4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.021225 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855"} err="failed to get container status \"4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855\": rpc error: code = NotFound desc = could not find container \"4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855\": container with ID starting with 4406b369f4545671a76cb8d08b48d04a10dadedb2bf9491fa0d9a0c94b09f855 not found: ID does not exist" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034090 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs4pb\" (UniqueName: \"kubernetes.io/projected/6a3529d7-ae55-48f8-9c32-392432471665-kube-api-access-vs4pb\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034122 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a3529d7-ae55-48f8-9c32-392432471665-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034137 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034148 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034158 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nf42\" (UniqueName: \"kubernetes.io/projected/dd0776fb-8edf-4508-98c3-5267fb06cd56-kube-api-access-7nf42\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034169 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd0776fb-8edf-4508-98c3-5267fb06cd56-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.034179 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a3529d7-ae55-48f8-9c32-392432471665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.282580 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.293286 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.302565 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.312621 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325109 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: E1126 06:44:06.325503 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-metadata" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325519 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-metadata" Nov 26 06:44:06 crc kubenswrapper[4775]: E1126 06:44:06.325555 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-log" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325562 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-log" Nov 26 06:44:06 crc kubenswrapper[4775]: E1126 06:44:06.325573 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0776fb-8edf-4508-98c3-5267fb06cd56" containerName="nova-cell1-novncproxy-novncproxy" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325581 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0776fb-8edf-4508-98c3-5267fb06cd56" containerName="nova-cell1-novncproxy-novncproxy" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325760 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-log" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325788 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3529d7-ae55-48f8-9c32-392432471665" containerName="nova-metadata-metadata" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.325800 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd0776fb-8edf-4508-98c3-5267fb06cd56" containerName="nova-cell1-novncproxy-novncproxy" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.326363 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.330455 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.330701 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.333436 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.359027 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.360615 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.363935 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.364506 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.378871 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.395824 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.440966 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441034 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vll4c\" (UniqueName: \"kubernetes.io/projected/c59f1995-e4fe-4355-89dc-52c4e3e13768-kube-api-access-vll4c\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441331 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441414 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c59f1995-e4fe-4355-89dc-52c4e3e13768-logs\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441582 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvmq8\" (UniqueName: \"kubernetes.io/projected/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-kube-api-access-nvmq8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441757 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.441857 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-config-data\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543565 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543634 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c59f1995-e4fe-4355-89dc-52c4e3e13768-logs\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543669 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543749 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvmq8\" (UniqueName: \"kubernetes.io/projected/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-kube-api-access-nvmq8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543797 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543833 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543862 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-config-data\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.543956 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vll4c\" (UniqueName: \"kubernetes.io/projected/c59f1995-e4fe-4355-89dc-52c4e3e13768-kube-api-access-vll4c\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.544267 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c59f1995-e4fe-4355-89dc-52c4e3e13768-logs\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.548666 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.550394 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.550446 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.550658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.551174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-config-data\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.552668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.557879 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.561184 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vll4c\" (UniqueName: \"kubernetes.io/projected/c59f1995-e4fe-4355-89dc-52c4e3e13768-kube-api-access-vll4c\") pod \"nova-metadata-0\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " pod="openstack/nova-metadata-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.570369 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvmq8\" (UniqueName: \"kubernetes.io/projected/c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211-kube-api-access-nvmq8\") pod \"nova-cell1-novncproxy-0\" (UID: \"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211\") " pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.655225 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:06 crc kubenswrapper[4775]: I1126 06:44:06.679588 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:07 crc kubenswrapper[4775]: W1126 06:44:07.165018 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7b39d9f_bad3_4e37_b5b0_f5aaa6b7c211.slice/crio-01fbd6217fbe4c8199e00f1e9193a772d4699e36df8bcc955b74430edffe62a0 WatchSource:0}: Error finding container 01fbd6217fbe4c8199e00f1e9193a772d4699e36df8bcc955b74430edffe62a0: Status 404 returned error can't find the container with id 01fbd6217fbe4c8199e00f1e9193a772d4699e36df8bcc955b74430edffe62a0 Nov 26 06:44:07 crc kubenswrapper[4775]: W1126 06:44:07.166129 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc59f1995_e4fe_4355_89dc_52c4e3e13768.slice/crio-88f299eae44d14624b299884f70903b778a2887a7f2d828f5de9fcf52140c815 WatchSource:0}: Error finding container 88f299eae44d14624b299884f70903b778a2887a7f2d828f5de9fcf52140c815: Status 404 returned error can't find the container with id 88f299eae44d14624b299884f70903b778a2887a7f2d828f5de9fcf52140c815 Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.171248 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.181283 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.360493 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3529d7-ae55-48f8-9c32-392432471665" path="/var/lib/kubelet/pods/6a3529d7-ae55-48f8-9c32-392432471665/volumes" Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.361865 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd0776fb-8edf-4508-98c3-5267fb06cd56" path="/var/lib/kubelet/pods/dd0776fb-8edf-4508-98c3-5267fb06cd56/volumes" Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.973256 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c59f1995-e4fe-4355-89dc-52c4e3e13768","Type":"ContainerStarted","Data":"5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b"} Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.973584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c59f1995-e4fe-4355-89dc-52c4e3e13768","Type":"ContainerStarted","Data":"279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de"} Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.973602 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c59f1995-e4fe-4355-89dc-52c4e3e13768","Type":"ContainerStarted","Data":"88f299eae44d14624b299884f70903b778a2887a7f2d828f5de9fcf52140c815"} Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.975983 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211","Type":"ContainerStarted","Data":"2da613328c443284411dac739edb5270e5b62672722ffd16c09be5df17b3b7cd"} Nov 26 06:44:07 crc kubenswrapper[4775]: I1126 06:44:07.976028 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211","Type":"ContainerStarted","Data":"01fbd6217fbe4c8199e00f1e9193a772d4699e36df8bcc955b74430edffe62a0"} Nov 26 06:44:08 crc kubenswrapper[4775]: I1126 06:44:08.000640 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.000619939 podStartE2EDuration="2.000619939s" podCreationTimestamp="2025-11-26 06:44:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:07.992385098 +0000 UTC m=+1131.353689050" watchObservedRunningTime="2025-11-26 06:44:08.000619939 +0000 UTC m=+1131.361923901" Nov 26 06:44:08 crc kubenswrapper[4775]: I1126 06:44:08.017889 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.017869749 podStartE2EDuration="2.017869749s" podCreationTimestamp="2025-11-26 06:44:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:08.010032147 +0000 UTC m=+1131.371336099" watchObservedRunningTime="2025-11-26 06:44:08.017869749 +0000 UTC m=+1131.379173701" Nov 26 06:44:09 crc kubenswrapper[4775]: I1126 06:44:09.158697 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 06:44:09 crc kubenswrapper[4775]: I1126 06:44:09.159485 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 06:44:09 crc kubenswrapper[4775]: I1126 06:44:09.161988 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 06:44:09 crc kubenswrapper[4775]: I1126 06:44:09.162733 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 06:44:09 crc kubenswrapper[4775]: I1126 06:44:09.995878 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 06:44:09 crc kubenswrapper[4775]: I1126 06:44:09.999570 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.201782 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-qkdkn"] Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.205823 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.229618 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-qkdkn"] Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.315573 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-config\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.315641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.315692 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.315725 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.315793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvdkf\" (UniqueName: \"kubernetes.io/projected/fcee54b8-a240-4821-82da-40e6649800d9-kube-api-access-hvdkf\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.315845 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.417180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-config\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.417484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.417526 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.417575 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.417655 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvdkf\" (UniqueName: \"kubernetes.io/projected/fcee54b8-a240-4821-82da-40e6649800d9-kube-api-access-hvdkf\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.417770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.418152 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-config\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.418930 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.419257 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.419359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.419805 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.436554 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvdkf\" (UniqueName: \"kubernetes.io/projected/fcee54b8-a240-4821-82da-40e6649800d9-kube-api-access-hvdkf\") pod \"dnsmasq-dns-59cf4bdb65-qkdkn\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:10 crc kubenswrapper[4775]: I1126 06:44:10.536449 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:11 crc kubenswrapper[4775]: I1126 06:44:11.034731 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-qkdkn"] Nov 26 06:44:11 crc kubenswrapper[4775]: I1126 06:44:11.656675 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:11 crc kubenswrapper[4775]: I1126 06:44:11.681143 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 06:44:11 crc kubenswrapper[4775]: I1126 06:44:11.681192 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.018239 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.018520 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-central-agent" containerID="cri-o://e45e03bd4bd0416b56133dafeb79943b4471f761aee19dbe23c9e15ed99c3c7b" gracePeriod=30 Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.018596 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="proxy-httpd" containerID="cri-o://ac8afae7aaf33e61197deda25d909984167c01878f67988ebda3ffc01644b29f" gracePeriod=30 Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.018639 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-notification-agent" containerID="cri-o://f82c48492922e8418e0c4d123c12575b533627198a2f55eb7cddc2f2848eda3a" gracePeriod=30 Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.018622 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="sg-core" containerID="cri-o://7f1d8222eb05820df1c0e0a107afb2e4a812df9650c61133fdc33692b3ee3ef6" gracePeriod=30 Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.020089 4775 generic.go:334] "Generic (PLEG): container finished" podID="fcee54b8-a240-4821-82da-40e6649800d9" containerID="bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e" exitCode=0 Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.021687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" event={"ID":"fcee54b8-a240-4821-82da-40e6649800d9","Type":"ContainerDied","Data":"bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e"} Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.021740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" event={"ID":"fcee54b8-a240-4821-82da-40e6649800d9","Type":"ContainerStarted","Data":"a08286bf6e21004210b12924cf89621db3aebc93bc8a2befa1bde5c3dce1f4fa"} Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.118706 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": read tcp 10.217.0.2:32918->10.217.0.194:3000: read: connection reset by peer" Nov 26 06:44:12 crc kubenswrapper[4775]: I1126 06:44:12.771489 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.038417 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" event={"ID":"fcee54b8-a240-4821-82da-40e6649800d9","Type":"ContainerStarted","Data":"0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962"} Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.038561 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041370 4775 generic.go:334] "Generic (PLEG): container finished" podID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerID="ac8afae7aaf33e61197deda25d909984167c01878f67988ebda3ffc01644b29f" exitCode=0 Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041408 4775 generic.go:334] "Generic (PLEG): container finished" podID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerID="7f1d8222eb05820df1c0e0a107afb2e4a812df9650c61133fdc33692b3ee3ef6" exitCode=2 Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041417 4775 generic.go:334] "Generic (PLEG): container finished" podID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerID="f82c48492922e8418e0c4d123c12575b533627198a2f55eb7cddc2f2848eda3a" exitCode=0 Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041426 4775 generic.go:334] "Generic (PLEG): container finished" podID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerID="e45e03bd4bd0416b56133dafeb79943b4471f761aee19dbe23c9e15ed99c3c7b" exitCode=0 Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041586 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-log" containerID="cri-o://cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299" gracePeriod=30 Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041664 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerDied","Data":"ac8afae7aaf33e61197deda25d909984167c01878f67988ebda3ffc01644b29f"} Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041694 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerDied","Data":"7f1d8222eb05820df1c0e0a107afb2e4a812df9650c61133fdc33692b3ee3ef6"} Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerDied","Data":"f82c48492922e8418e0c4d123c12575b533627198a2f55eb7cddc2f2848eda3a"} Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041719 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerDied","Data":"e45e03bd4bd0416b56133dafeb79943b4471f761aee19dbe23c9e15ed99c3c7b"} Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.041780 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-api" containerID="cri-o://e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b" gracePeriod=30 Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.064497 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" podStartSLOduration=3.064481338 podStartE2EDuration="3.064481338s" podCreationTimestamp="2025-11-26 06:44:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:13.061232023 +0000 UTC m=+1136.422535985" watchObservedRunningTime="2025-11-26 06:44:13.064481338 +0000 UTC m=+1136.425785280" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.246058 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.290806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2knrs\" (UniqueName: \"kubernetes.io/projected/74c15881-a4e4-490e-81dc-d6acfae420ab-kube-api-access-2knrs\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.290872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-sg-core-conf-yaml\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.290894 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-scripts\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.290973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-run-httpd\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.290995 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-ceilometer-tls-certs\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.291026 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-config-data\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.291154 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-log-httpd\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.291280 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-combined-ca-bundle\") pod \"74c15881-a4e4-490e-81dc-d6acfae420ab\" (UID: \"74c15881-a4e4-490e-81dc-d6acfae420ab\") " Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.291393 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.291668 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.295027 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.298854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-scripts" (OuterVolumeSpecName: "scripts") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.303421 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c15881-a4e4-490e-81dc-d6acfae420ab-kube-api-access-2knrs" (OuterVolumeSpecName: "kube-api-access-2knrs") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "kube-api-access-2knrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.330488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.364140 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.382324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.393625 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2knrs\" (UniqueName: \"kubernetes.io/projected/74c15881-a4e4-490e-81dc-d6acfae420ab-kube-api-access-2knrs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.393649 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.393660 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.393669 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.393677 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/74c15881-a4e4-490e-81dc-d6acfae420ab-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.393686 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.420932 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-config-data" (OuterVolumeSpecName: "config-data") pod "74c15881-a4e4-490e-81dc-d6acfae420ab" (UID: "74c15881-a4e4-490e-81dc-d6acfae420ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:13 crc kubenswrapper[4775]: I1126 06:44:13.495548 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74c15881-a4e4-490e-81dc-d6acfae420ab-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.049882 4775 generic.go:334] "Generic (PLEG): container finished" podID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerID="cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299" exitCode=143 Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.049952 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c60149bf-8618-4ecd-9cd7-80d59881a7b6","Type":"ContainerDied","Data":"cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299"} Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.053001 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"74c15881-a4e4-490e-81dc-d6acfae420ab","Type":"ContainerDied","Data":"35d6f056953144549723f9407089092caefce6db649775db7c3f713b63630ff5"} Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.053033 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.053035 4775 scope.go:117] "RemoveContainer" containerID="ac8afae7aaf33e61197deda25d909984167c01878f67988ebda3ffc01644b29f" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.092038 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.093779 4775 scope.go:117] "RemoveContainer" containerID="7f1d8222eb05820df1c0e0a107afb2e4a812df9650c61133fdc33692b3ee3ef6" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.127998 4775 scope.go:117] "RemoveContainer" containerID="f82c48492922e8418e0c4d123c12575b533627198a2f55eb7cddc2f2848eda3a" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.128289 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.172186 4775 scope.go:117] "RemoveContainer" containerID="e45e03bd4bd0416b56133dafeb79943b4471f761aee19dbe23c9e15ed99c3c7b" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.173960 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:44:14 crc kubenswrapper[4775]: E1126 06:44:14.174818 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-central-agent" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.174928 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-central-agent" Nov 26 06:44:14 crc kubenswrapper[4775]: E1126 06:44:14.175020 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-notification-agent" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.175092 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-notification-agent" Nov 26 06:44:14 crc kubenswrapper[4775]: E1126 06:44:14.175166 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="sg-core" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.175241 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="sg-core" Nov 26 06:44:14 crc kubenswrapper[4775]: E1126 06:44:14.175323 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="proxy-httpd" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.175409 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="proxy-httpd" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.175847 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="sg-core" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.176102 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-central-agent" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.176187 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="ceilometer-notification-agent" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.176269 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" containerName="proxy-httpd" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.178910 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.180403 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.180988 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.181388 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.184403 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.319343 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.319629 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-config-data\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.319744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtvdg\" (UniqueName: \"kubernetes.io/projected/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-kube-api-access-xtvdg\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.319848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-run-httpd\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.319965 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.320033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-scripts\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.320126 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.320242 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-log-httpd\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422104 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-run-httpd\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422278 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-scripts\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422316 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-log-httpd\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422395 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422438 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-config-data\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.422474 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtvdg\" (UniqueName: \"kubernetes.io/projected/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-kube-api-access-xtvdg\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.423241 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-run-httpd\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.423668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-log-httpd\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.427791 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-scripts\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.428377 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.429238 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.429650 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-config-data\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.435851 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.442497 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtvdg\" (UniqueName: \"kubernetes.io/projected/9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7-kube-api-access-xtvdg\") pod \"ceilometer-0\" (UID: \"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7\") " pod="openstack/ceilometer-0" Nov 26 06:44:14 crc kubenswrapper[4775]: I1126 06:44:14.504064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 26 06:44:15 crc kubenswrapper[4775]: I1126 06:44:15.013419 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 26 06:44:15 crc kubenswrapper[4775]: I1126 06:44:15.063686 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7","Type":"ContainerStarted","Data":"4104cbf069d6f1fab527704207a56b0ded8df884df1c0769b105477f5cab5c8b"} Nov 26 06:44:15 crc kubenswrapper[4775]: I1126 06:44:15.343672 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c15881-a4e4-490e-81dc-d6acfae420ab" path="/var/lib/kubelet/pods/74c15881-a4e4-490e-81dc-d6acfae420ab/volumes" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.084846 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7","Type":"ContainerStarted","Data":"3713446345e1ea59b6e059d6a4059fc045fca46f56ffd9fc3ccd92c48a02a31b"} Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.612002 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.656940 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.680852 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.680914 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.685254 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.807022 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-combined-ca-bundle\") pod \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.807153 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44fvx\" (UniqueName: \"kubernetes.io/projected/c60149bf-8618-4ecd-9cd7-80d59881a7b6-kube-api-access-44fvx\") pod \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.807262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60149bf-8618-4ecd-9cd7-80d59881a7b6-logs\") pod \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.807287 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-config-data\") pod \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\" (UID: \"c60149bf-8618-4ecd-9cd7-80d59881a7b6\") " Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.808600 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c60149bf-8618-4ecd-9cd7-80d59881a7b6-logs" (OuterVolumeSpecName: "logs") pod "c60149bf-8618-4ecd-9cd7-80d59881a7b6" (UID: "c60149bf-8618-4ecd-9cd7-80d59881a7b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.817742 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c60149bf-8618-4ecd-9cd7-80d59881a7b6-kube-api-access-44fvx" (OuterVolumeSpecName: "kube-api-access-44fvx") pod "c60149bf-8618-4ecd-9cd7-80d59881a7b6" (UID: "c60149bf-8618-4ecd-9cd7-80d59881a7b6"). InnerVolumeSpecName "kube-api-access-44fvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.839831 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-config-data" (OuterVolumeSpecName: "config-data") pod "c60149bf-8618-4ecd-9cd7-80d59881a7b6" (UID: "c60149bf-8618-4ecd-9cd7-80d59881a7b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.858853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c60149bf-8618-4ecd-9cd7-80d59881a7b6" (UID: "c60149bf-8618-4ecd-9cd7-80d59881a7b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.909788 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c60149bf-8618-4ecd-9cd7-80d59881a7b6-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.909826 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.909836 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c60149bf-8618-4ecd-9cd7-80d59881a7b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:16 crc kubenswrapper[4775]: I1126 06:44:16.909849 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44fvx\" (UniqueName: \"kubernetes.io/projected/c60149bf-8618-4ecd-9cd7-80d59881a7b6-kube-api-access-44fvx\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.114204 4775 generic.go:334] "Generic (PLEG): container finished" podID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerID="e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b" exitCode=0 Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.114440 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.114455 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c60149bf-8618-4ecd-9cd7-80d59881a7b6","Type":"ContainerDied","Data":"e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b"} Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.114871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c60149bf-8618-4ecd-9cd7-80d59881a7b6","Type":"ContainerDied","Data":"9b653d5114454fb6073a515b6bdf6a3fd6dadbede69feb91b1fff970ea02c585"} Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.114922 4775 scope.go:117] "RemoveContainer" containerID="e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.122017 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7","Type":"ContainerStarted","Data":"ab3d7c842a68cc3ed49105bb45d715fe35b94c89a6e5ec2bc52dbdda32effdea"} Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.149766 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.149903 4775 scope.go:117] "RemoveContainer" containerID="cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.164968 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.177784 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.209798 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:17 crc kubenswrapper[4775]: E1126 06:44:17.210295 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-log" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.210313 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-log" Nov 26 06:44:17 crc kubenswrapper[4775]: E1126 06:44:17.210349 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-api" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.210357 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-api" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.210606 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-log" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.210631 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" containerName="nova-api-api" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.211894 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.217084 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.217235 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.217682 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.218704 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f0fe4e3-3dda-4a22-932f-1389707a264c-logs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.218751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhx48\" (UniqueName: \"kubernetes.io/projected/7f0fe4e3-3dda-4a22-932f-1389707a264c-kube-api-access-zhx48\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.218790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.218827 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.218852 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.218997 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-config-data\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.219135 4775 scope.go:117] "RemoveContainer" containerID="e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b" Nov 26 06:44:17 crc kubenswrapper[4775]: E1126 06:44:17.222804 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b\": container with ID starting with e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b not found: ID does not exist" containerID="e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.222854 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b"} err="failed to get container status \"e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b\": rpc error: code = NotFound desc = could not find container \"e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b\": container with ID starting with e2d363aa37467e1b863ba352092c703d0df014ab324cd4ebd13d6cfe9b5f328b not found: ID does not exist" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.222900 4775 scope.go:117] "RemoveContainer" containerID="cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.224038 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:17 crc kubenswrapper[4775]: E1126 06:44:17.224254 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299\": container with ID starting with cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299 not found: ID does not exist" containerID="cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.224284 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299"} err="failed to get container status \"cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299\": rpc error: code = NotFound desc = could not find container \"cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299\": container with ID starting with cf12491da12da272dbc4540509b14ccc5d5457ccf65002cb41ec60d9fa677299 not found: ID does not exist" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f0fe4e3-3dda-4a22-932f-1389707a264c-logs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhx48\" (UniqueName: \"kubernetes.io/projected/7f0fe4e3-3dda-4a22-932f-1389707a264c-kube-api-access-zhx48\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321150 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321179 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-config-data\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.321559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f0fe4e3-3dda-4a22-932f-1389707a264c-logs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.323598 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.323886 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.324002 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.332521 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.335986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.338561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-config-data\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.339810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.350484 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c60149bf-8618-4ecd-9cd7-80d59881a7b6" path="/var/lib/kubelet/pods/c60149bf-8618-4ecd-9cd7-80d59881a7b6/volumes" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.350838 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhx48\" (UniqueName: \"kubernetes.io/projected/7f0fe4e3-3dda-4a22-932f-1389707a264c-kube-api-access-zhx48\") pod \"nova-api-0\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.479144 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-96s25"] Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.480605 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.484406 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.484932 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.493260 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-96s25"] Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.532300 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.626977 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-config-data\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.627153 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-scripts\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.627260 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtzvl\" (UniqueName: \"kubernetes.io/projected/8ba23799-9d71-495a-a716-56dfca1804a2-kube-api-access-xtzvl\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.627338 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.697915 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.697907 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.196:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.728760 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-scripts\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.728856 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtzvl\" (UniqueName: \"kubernetes.io/projected/8ba23799-9d71-495a-a716-56dfca1804a2-kube-api-access-xtzvl\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.728901 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.728974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-config-data\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.736228 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-scripts\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.755867 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-config-data\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.755948 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.759245 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtzvl\" (UniqueName: \"kubernetes.io/projected/8ba23799-9d71-495a-a716-56dfca1804a2-kube-api-access-xtzvl\") pod \"nova-cell1-cell-mapping-96s25\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:17 crc kubenswrapper[4775]: I1126 06:44:17.797807 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:18 crc kubenswrapper[4775]: I1126 06:44:18.007356 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:18 crc kubenswrapper[4775]: I1126 06:44:18.131318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f0fe4e3-3dda-4a22-932f-1389707a264c","Type":"ContainerStarted","Data":"3791d7195f4834fb2a6ae7b06ba03d4f79bae060a4c01d5b5eb49583e17e4abd"} Nov 26 06:44:18 crc kubenswrapper[4775]: I1126 06:44:18.148969 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7","Type":"ContainerStarted","Data":"4def7b2bbe3d37b7f833163ea531a73f84c85c94ff15c7116362fbe903199969"} Nov 26 06:44:18 crc kubenswrapper[4775]: I1126 06:44:18.251321 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-96s25"] Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.161760 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7","Type":"ContainerStarted","Data":"bf572f96ef53fa86cd4f1189918156ebff1241d8f4519274f33d1267e93fea9e"} Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.161881 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.165011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96s25" event={"ID":"8ba23799-9d71-495a-a716-56dfca1804a2","Type":"ContainerStarted","Data":"f140ad268db422358cf37f051ef0d48fa16a5f6a8ef3d211110790f4df383a28"} Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.165042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96s25" event={"ID":"8ba23799-9d71-495a-a716-56dfca1804a2","Type":"ContainerStarted","Data":"129916e11ac8928ba481f33435887d2200088a49eb41902b46772b6b341e0aa6"} Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.167336 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f0fe4e3-3dda-4a22-932f-1389707a264c","Type":"ContainerStarted","Data":"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15"} Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.167361 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f0fe4e3-3dda-4a22-932f-1389707a264c","Type":"ContainerStarted","Data":"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d"} Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.194565 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.954803925 podStartE2EDuration="5.194536529s" podCreationTimestamp="2025-11-26 06:44:14 +0000 UTC" firstStartedPulling="2025-11-26 06:44:15.012682968 +0000 UTC m=+1138.373986940" lastFinishedPulling="2025-11-26 06:44:18.252415592 +0000 UTC m=+1141.613719544" observedRunningTime="2025-11-26 06:44:19.19342236 +0000 UTC m=+1142.554726332" watchObservedRunningTime="2025-11-26 06:44:19.194536529 +0000 UTC m=+1142.555840491" Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.233045 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.233023438 podStartE2EDuration="2.233023438s" podCreationTimestamp="2025-11-26 06:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:19.219340836 +0000 UTC m=+1142.580644808" watchObservedRunningTime="2025-11-26 06:44:19.233023438 +0000 UTC m=+1142.594327430" Nov 26 06:44:19 crc kubenswrapper[4775]: I1126 06:44:19.250910 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-96s25" podStartSLOduration=2.250883221 podStartE2EDuration="2.250883221s" podCreationTimestamp="2025-11-26 06:44:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:19.243894926 +0000 UTC m=+1142.605198888" watchObservedRunningTime="2025-11-26 06:44:19.250883221 +0000 UTC m=+1142.612187183" Nov 26 06:44:20 crc kubenswrapper[4775]: I1126 06:44:20.537854 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:44:20 crc kubenswrapper[4775]: I1126 06:44:20.604592 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-7q5vm"] Nov 26 06:44:20 crc kubenswrapper[4775]: I1126 06:44:20.604851 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerName="dnsmasq-dns" containerID="cri-o://20a19d0b0eb11331e934143d7cff3572c583c8581e65550c8b8a42cce159863a" gracePeriod=10 Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.190279 4775 generic.go:334] "Generic (PLEG): container finished" podID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerID="20a19d0b0eb11331e934143d7cff3572c583c8581e65550c8b8a42cce159863a" exitCode=0 Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.190489 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" event={"ID":"ef02f1ef-024a-4c85-826f-7b0c4749dbdd","Type":"ContainerDied","Data":"20a19d0b0eb11331e934143d7cff3572c583c8581e65550c8b8a42cce159863a"} Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.190571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" event={"ID":"ef02f1ef-024a-4c85-826f-7b0c4749dbdd","Type":"ContainerDied","Data":"512fdbe0f2201de777a9ed7a34e2d5d03f0e05267cce70455224d87aded20077"} Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.190587 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="512fdbe0f2201de777a9ed7a34e2d5d03f0e05267cce70455224d87aded20077" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.207448 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.305144 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-sb\") pod \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.305210 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcpzf\" (UniqueName: \"kubernetes.io/projected/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-kube-api-access-lcpzf\") pod \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.305327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-config\") pod \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.305388 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-swift-storage-0\") pod \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.305415 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-nb\") pod \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.305483 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-svc\") pod \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\" (UID: \"ef02f1ef-024a-4c85-826f-7b0c4749dbdd\") " Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.316847 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-kube-api-access-lcpzf" (OuterVolumeSpecName: "kube-api-access-lcpzf") pod "ef02f1ef-024a-4c85-826f-7b0c4749dbdd" (UID: "ef02f1ef-024a-4c85-826f-7b0c4749dbdd"). InnerVolumeSpecName "kube-api-access-lcpzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.363343 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-config" (OuterVolumeSpecName: "config") pod "ef02f1ef-024a-4c85-826f-7b0c4749dbdd" (UID: "ef02f1ef-024a-4c85-826f-7b0c4749dbdd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.385786 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef02f1ef-024a-4c85-826f-7b0c4749dbdd" (UID: "ef02f1ef-024a-4c85-826f-7b0c4749dbdd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.386354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef02f1ef-024a-4c85-826f-7b0c4749dbdd" (UID: "ef02f1ef-024a-4c85-826f-7b0c4749dbdd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.387970 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ef02f1ef-024a-4c85-826f-7b0c4749dbdd" (UID: "ef02f1ef-024a-4c85-826f-7b0c4749dbdd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.404734 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef02f1ef-024a-4c85-826f-7b0c4749dbdd" (UID: "ef02f1ef-024a-4c85-826f-7b0c4749dbdd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.407547 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.407568 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.407577 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.407586 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.407595 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:21 crc kubenswrapper[4775]: I1126 06:44:21.407602 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcpzf\" (UniqueName: \"kubernetes.io/projected/ef02f1ef-024a-4c85-826f-7b0c4749dbdd-kube-api-access-lcpzf\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:22 crc kubenswrapper[4775]: I1126 06:44:22.198648 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-7q5vm" Nov 26 06:44:22 crc kubenswrapper[4775]: I1126 06:44:22.233442 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-7q5vm"] Nov 26 06:44:22 crc kubenswrapper[4775]: I1126 06:44:22.247410 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-7q5vm"] Nov 26 06:44:23 crc kubenswrapper[4775]: I1126 06:44:23.211800 4775 generic.go:334] "Generic (PLEG): container finished" podID="8ba23799-9d71-495a-a716-56dfca1804a2" containerID="f140ad268db422358cf37f051ef0d48fa16a5f6a8ef3d211110790f4df383a28" exitCode=0 Nov 26 06:44:23 crc kubenswrapper[4775]: I1126 06:44:23.211894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96s25" event={"ID":"8ba23799-9d71-495a-a716-56dfca1804a2","Type":"ContainerDied","Data":"f140ad268db422358cf37f051ef0d48fa16a5f6a8ef3d211110790f4df383a28"} Nov 26 06:44:23 crc kubenswrapper[4775]: I1126 06:44:23.341416 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" path="/var/lib/kubelet/pods/ef02f1ef-024a-4c85-826f-7b0c4749dbdd/volumes" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.611044 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.676699 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-scripts\") pod \"8ba23799-9d71-495a-a716-56dfca1804a2\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.676878 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-combined-ca-bundle\") pod \"8ba23799-9d71-495a-a716-56dfca1804a2\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.676925 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-config-data\") pod \"8ba23799-9d71-495a-a716-56dfca1804a2\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.676972 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtzvl\" (UniqueName: \"kubernetes.io/projected/8ba23799-9d71-495a-a716-56dfca1804a2-kube-api-access-xtzvl\") pod \"8ba23799-9d71-495a-a716-56dfca1804a2\" (UID: \"8ba23799-9d71-495a-a716-56dfca1804a2\") " Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.682451 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ba23799-9d71-495a-a716-56dfca1804a2-kube-api-access-xtzvl" (OuterVolumeSpecName: "kube-api-access-xtzvl") pod "8ba23799-9d71-495a-a716-56dfca1804a2" (UID: "8ba23799-9d71-495a-a716-56dfca1804a2"). InnerVolumeSpecName "kube-api-access-xtzvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.684214 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-scripts" (OuterVolumeSpecName: "scripts") pod "8ba23799-9d71-495a-a716-56dfca1804a2" (UID: "8ba23799-9d71-495a-a716-56dfca1804a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.710174 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-config-data" (OuterVolumeSpecName: "config-data") pod "8ba23799-9d71-495a-a716-56dfca1804a2" (UID: "8ba23799-9d71-495a-a716-56dfca1804a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.713906 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ba23799-9d71-495a-a716-56dfca1804a2" (UID: "8ba23799-9d71-495a-a716-56dfca1804a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.779541 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.779571 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.779583 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ba23799-9d71-495a-a716-56dfca1804a2-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:24 crc kubenswrapper[4775]: I1126 06:44:24.779593 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtzvl\" (UniqueName: \"kubernetes.io/projected/8ba23799-9d71-495a-a716-56dfca1804a2-kube-api-access-xtzvl\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.242129 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-96s25" event={"ID":"8ba23799-9d71-495a-a716-56dfca1804a2","Type":"ContainerDied","Data":"129916e11ac8928ba481f33435887d2200088a49eb41902b46772b6b341e0aa6"} Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.242181 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="129916e11ac8928ba481f33435887d2200088a49eb41902b46772b6b341e0aa6" Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.242219 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-96s25" Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.438585 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.438898 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" containerName="nova-scheduler-scheduler" containerID="cri-o://57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae" gracePeriod=30 Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.459171 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.459519 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-log" containerID="cri-o://d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d" gracePeriod=30 Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.459655 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-api" containerID="cri-o://6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15" gracePeriod=30 Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.512737 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.512994 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-log" containerID="cri-o://279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de" gracePeriod=30 Nov 26 06:44:25 crc kubenswrapper[4775]: I1126 06:44:25.513112 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-metadata" containerID="cri-o://5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b" gracePeriod=30 Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.016105 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.109177 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhx48\" (UniqueName: \"kubernetes.io/projected/7f0fe4e3-3dda-4a22-932f-1389707a264c-kube-api-access-zhx48\") pod \"7f0fe4e3-3dda-4a22-932f-1389707a264c\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.109233 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f0fe4e3-3dda-4a22-932f-1389707a264c-logs\") pod \"7f0fe4e3-3dda-4a22-932f-1389707a264c\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.109271 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-combined-ca-bundle\") pod \"7f0fe4e3-3dda-4a22-932f-1389707a264c\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.109378 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-internal-tls-certs\") pod \"7f0fe4e3-3dda-4a22-932f-1389707a264c\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.109447 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-config-data\") pod \"7f0fe4e3-3dda-4a22-932f-1389707a264c\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.109473 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-public-tls-certs\") pod \"7f0fe4e3-3dda-4a22-932f-1389707a264c\" (UID: \"7f0fe4e3-3dda-4a22-932f-1389707a264c\") " Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.110398 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f0fe4e3-3dda-4a22-932f-1389707a264c-logs" (OuterVolumeSpecName: "logs") pod "7f0fe4e3-3dda-4a22-932f-1389707a264c" (UID: "7f0fe4e3-3dda-4a22-932f-1389707a264c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.117884 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f0fe4e3-3dda-4a22-932f-1389707a264c-kube-api-access-zhx48" (OuterVolumeSpecName: "kube-api-access-zhx48") pod "7f0fe4e3-3dda-4a22-932f-1389707a264c" (UID: "7f0fe4e3-3dda-4a22-932f-1389707a264c"). InnerVolumeSpecName "kube-api-access-zhx48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.146614 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-config-data" (OuterVolumeSpecName: "config-data") pod "7f0fe4e3-3dda-4a22-932f-1389707a264c" (UID: "7f0fe4e3-3dda-4a22-932f-1389707a264c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.162336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7f0fe4e3-3dda-4a22-932f-1389707a264c" (UID: "7f0fe4e3-3dda-4a22-932f-1389707a264c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.165342 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f0fe4e3-3dda-4a22-932f-1389707a264c" (UID: "7f0fe4e3-3dda-4a22-932f-1389707a264c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.165763 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7f0fe4e3-3dda-4a22-932f-1389707a264c" (UID: "7f0fe4e3-3dda-4a22-932f-1389707a264c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.211733 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhx48\" (UniqueName: \"kubernetes.io/projected/7f0fe4e3-3dda-4a22-932f-1389707a264c-kube-api-access-zhx48\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.211978 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f0fe4e3-3dda-4a22-932f-1389707a264c-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.212075 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.212180 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.212249 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.212308 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f0fe4e3-3dda-4a22-932f-1389707a264c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.252313 4775 generic.go:334] "Generic (PLEG): container finished" podID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerID="279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de" exitCode=143 Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.252728 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c59f1995-e4fe-4355-89dc-52c4e3e13768","Type":"ContainerDied","Data":"279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de"} Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.256688 4775 generic.go:334] "Generic (PLEG): container finished" podID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerID="6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15" exitCode=0 Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.256942 4775 generic.go:334] "Generic (PLEG): container finished" podID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerID="d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d" exitCode=143 Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.256762 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f0fe4e3-3dda-4a22-932f-1389707a264c","Type":"ContainerDied","Data":"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15"} Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.257218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f0fe4e3-3dda-4a22-932f-1389707a264c","Type":"ContainerDied","Data":"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d"} Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.257359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7f0fe4e3-3dda-4a22-932f-1389707a264c","Type":"ContainerDied","Data":"3791d7195f4834fb2a6ae7b06ba03d4f79bae060a4c01d5b5eb49583e17e4abd"} Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.256750 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.257278 4775 scope.go:117] "RemoveContainer" containerID="6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.300569 4775 scope.go:117] "RemoveContainer" containerID="d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.314790 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.336691 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.348396 4775 scope.go:117] "RemoveContainer" containerID="6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.349016 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15\": container with ID starting with 6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15 not found: ID does not exist" containerID="6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349053 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15"} err="failed to get container status \"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15\": rpc error: code = NotFound desc = could not find container \"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15\": container with ID starting with 6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15 not found: ID does not exist" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349075 4775 scope.go:117] "RemoveContainer" containerID="d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.349413 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d\": container with ID starting with d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d not found: ID does not exist" containerID="d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349437 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d"} err="failed to get container status \"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d\": rpc error: code = NotFound desc = could not find container \"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d\": container with ID starting with d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d not found: ID does not exist" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349450 4775 scope.go:117] "RemoveContainer" containerID="6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349779 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15"} err="failed to get container status \"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15\": rpc error: code = NotFound desc = could not find container \"6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15\": container with ID starting with 6aeb0b0b9b42b5a126a58d038c0e186385a12c36d52c11967ac4bea58cae5c15 not found: ID does not exist" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349796 4775 scope.go:117] "RemoveContainer" containerID="d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.349957 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350142 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d"} err="failed to get container status \"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d\": rpc error: code = NotFound desc = could not find container \"d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d\": container with ID starting with d1a056e357087a85f0a0cca452913d56ea29c7b7cfe5899ce355a1e948788f0d not found: ID does not exist" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.350342 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-api" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350358 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-api" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.350381 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba23799-9d71-495a-a716-56dfca1804a2" containerName="nova-manage" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350388 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba23799-9d71-495a-a716-56dfca1804a2" containerName="nova-manage" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.350412 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerName="dnsmasq-dns" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350418 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerName="dnsmasq-dns" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.350430 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerName="init" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350435 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerName="init" Nov 26 06:44:26 crc kubenswrapper[4775]: E1126 06:44:26.350452 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-log" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350457 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-log" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350813 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-api" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350841 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef02f1ef-024a-4c85-826f-7b0c4749dbdd" containerName="dnsmasq-dns" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350861 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" containerName="nova-api-log" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.350877 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba23799-9d71-495a-a716-56dfca1804a2" containerName="nova-manage" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.351899 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.354214 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.354367 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.364079 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.366671 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.419318 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vqmn\" (UniqueName: \"kubernetes.io/projected/09a96730-d166-4111-b88a-b2a51ca06b5a-kube-api-access-7vqmn\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.419381 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-public-tls-certs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.419409 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-config-data\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.419463 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.419524 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.419552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09a96730-d166-4111-b88a-b2a51ca06b5a-logs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521030 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521093 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09a96730-d166-4111-b88a-b2a51ca06b5a-logs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521229 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vqmn\" (UniqueName: \"kubernetes.io/projected/09a96730-d166-4111-b88a-b2a51ca06b5a-kube-api-access-7vqmn\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521268 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-public-tls-certs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521312 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-config-data\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521345 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.521938 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09a96730-d166-4111-b88a-b2a51ca06b5a-logs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.524673 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-config-data\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.525088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-public-tls-certs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.525268 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.525499 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09a96730-d166-4111-b88a-b2a51ca06b5a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.541901 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vqmn\" (UniqueName: \"kubernetes.io/projected/09a96730-d166-4111-b88a-b2a51ca06b5a-kube-api-access-7vqmn\") pod \"nova-api-0\" (UID: \"09a96730-d166-4111-b88a-b2a51ca06b5a\") " pod="openstack/nova-api-0" Nov 26 06:44:26 crc kubenswrapper[4775]: I1126 06:44:26.681371 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.026315 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.265893 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.269943 4775 generic.go:334] "Generic (PLEG): container finished" podID="c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" containerID="57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae" exitCode=0 Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.270011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00","Type":"ContainerDied","Data":"57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae"} Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.270040 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00","Type":"ContainerDied","Data":"91f2ddb7b2d3843a1b92f31471723eaa1a387659388591da37acec746e32e853"} Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.270057 4775 scope.go:117] "RemoveContainer" containerID="57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.270180 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.275388 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"09a96730-d166-4111-b88a-b2a51ca06b5a","Type":"ContainerStarted","Data":"0cbb39e01395c4b28dffade82d45da792370ca25ea03efcbd3682ada3b6ffbe3"} Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.275430 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"09a96730-d166-4111-b88a-b2a51ca06b5a","Type":"ContainerStarted","Data":"fdbdd951e8931f3a73fbc33efae77cbe5cbb4af87aa014094f4fd84b034f3368"} Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.291320 4775 scope.go:117] "RemoveContainer" containerID="57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae" Nov 26 06:44:27 crc kubenswrapper[4775]: E1126 06:44:27.293039 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae\": container with ID starting with 57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae not found: ID does not exist" containerID="57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.293101 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae"} err="failed to get container status \"57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae\": rpc error: code = NotFound desc = could not find container \"57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae\": container with ID starting with 57ca393bbdd4982055e0b24a08e9a4d71aa0c7990f122618c4071204adaab4ae not found: ID does not exist" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.346404 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f0fe4e3-3dda-4a22-932f-1389707a264c" path="/var/lib/kubelet/pods/7f0fe4e3-3dda-4a22-932f-1389707a264c/volumes" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.349880 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-combined-ca-bundle\") pod \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.349990 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9s74\" (UniqueName: \"kubernetes.io/projected/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-kube-api-access-z9s74\") pod \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.350051 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-config-data\") pod \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\" (UID: \"c8ab86ff-2e8a-424f-9c94-6522b7e6ab00\") " Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.355386 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-kube-api-access-z9s74" (OuterVolumeSpecName: "kube-api-access-z9s74") pod "c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" (UID: "c8ab86ff-2e8a-424f-9c94-6522b7e6ab00"). InnerVolumeSpecName "kube-api-access-z9s74". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.380228 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" (UID: "c8ab86ff-2e8a-424f-9c94-6522b7e6ab00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.395167 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-config-data" (OuterVolumeSpecName: "config-data") pod "c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" (UID: "c8ab86ff-2e8a-424f-9c94-6522b7e6ab00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.454558 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.454590 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9s74\" (UniqueName: \"kubernetes.io/projected/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-kube-api-access-z9s74\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.454602 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.674086 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.688706 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.714265 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:44:27 crc kubenswrapper[4775]: E1126 06:44:27.714655 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" containerName="nova-scheduler-scheduler" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.714672 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" containerName="nova-scheduler-scheduler" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.714892 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" containerName="nova-scheduler-scheduler" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.720138 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.722883 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.729178 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.760593 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8d44208-4c24-4835-89f6-a471ce50ddae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.760654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8d44208-4c24-4835-89f6-a471ce50ddae-config-data\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.760745 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jhcg\" (UniqueName: \"kubernetes.io/projected/e8d44208-4c24-4835-89f6-a471ce50ddae-kube-api-access-9jhcg\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.863183 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8d44208-4c24-4835-89f6-a471ce50ddae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.863686 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8d44208-4c24-4835-89f6-a471ce50ddae-config-data\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.864034 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jhcg\" (UniqueName: \"kubernetes.io/projected/e8d44208-4c24-4835-89f6-a471ce50ddae-kube-api-access-9jhcg\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.869540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8d44208-4c24-4835-89f6-a471ce50ddae-config-data\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.870062 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8d44208-4c24-4835-89f6-a471ce50ddae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:27 crc kubenswrapper[4775]: I1126 06:44:27.883836 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jhcg\" (UniqueName: \"kubernetes.io/projected/e8d44208-4c24-4835-89f6-a471ce50ddae-kube-api-access-9jhcg\") pod \"nova-scheduler-0\" (UID: \"e8d44208-4c24-4835-89f6-a471ce50ddae\") " pod="openstack/nova-scheduler-0" Nov 26 06:44:28 crc kubenswrapper[4775]: I1126 06:44:28.050247 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 26 06:44:28 crc kubenswrapper[4775]: I1126 06:44:28.288930 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"09a96730-d166-4111-b88a-b2a51ca06b5a","Type":"ContainerStarted","Data":"aa91c90fb783151cc06dca31588289d7588c94fad52a7792bf2e77f76de6ca3b"} Nov 26 06:44:28 crc kubenswrapper[4775]: I1126 06:44:28.320432 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.320400942 podStartE2EDuration="2.320400942s" podCreationTimestamp="2025-11-26 06:44:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:28.306519145 +0000 UTC m=+1151.667823097" watchObservedRunningTime="2025-11-26 06:44:28.320400942 +0000 UTC m=+1151.681704894" Nov 26 06:44:28 crc kubenswrapper[4775]: I1126 06:44:28.550909 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 26 06:44:28 crc kubenswrapper[4775]: W1126 06:44:28.558250 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8d44208_4c24_4835_89f6_a471ce50ddae.slice/crio-7654588dab175a0903d1038819ffbd5b76517e597ab59cad1d7324d2d6532ddf WatchSource:0}: Error finding container 7654588dab175a0903d1038819ffbd5b76517e597ab59cad1d7324d2d6532ddf: Status 404 returned error can't find the container with id 7654588dab175a0903d1038819ffbd5b76517e597ab59cad1d7324d2d6532ddf Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.082161 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.198198 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-nova-metadata-tls-certs\") pod \"c59f1995-e4fe-4355-89dc-52c4e3e13768\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.198286 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c59f1995-e4fe-4355-89dc-52c4e3e13768-logs\") pod \"c59f1995-e4fe-4355-89dc-52c4e3e13768\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.198440 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-combined-ca-bundle\") pod \"c59f1995-e4fe-4355-89dc-52c4e3e13768\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.198491 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vll4c\" (UniqueName: \"kubernetes.io/projected/c59f1995-e4fe-4355-89dc-52c4e3e13768-kube-api-access-vll4c\") pod \"c59f1995-e4fe-4355-89dc-52c4e3e13768\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.198601 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-config-data\") pod \"c59f1995-e4fe-4355-89dc-52c4e3e13768\" (UID: \"c59f1995-e4fe-4355-89dc-52c4e3e13768\") " Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.200276 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c59f1995-e4fe-4355-89dc-52c4e3e13768-logs" (OuterVolumeSpecName: "logs") pod "c59f1995-e4fe-4355-89dc-52c4e3e13768" (UID: "c59f1995-e4fe-4355-89dc-52c4e3e13768"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.203398 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c59f1995-e4fe-4355-89dc-52c4e3e13768-kube-api-access-vll4c" (OuterVolumeSpecName: "kube-api-access-vll4c") pod "c59f1995-e4fe-4355-89dc-52c4e3e13768" (UID: "c59f1995-e4fe-4355-89dc-52c4e3e13768"). InnerVolumeSpecName "kube-api-access-vll4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.235305 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c59f1995-e4fe-4355-89dc-52c4e3e13768" (UID: "c59f1995-e4fe-4355-89dc-52c4e3e13768"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.239556 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-config-data" (OuterVolumeSpecName: "config-data") pod "c59f1995-e4fe-4355-89dc-52c4e3e13768" (UID: "c59f1995-e4fe-4355-89dc-52c4e3e13768"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.254053 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c59f1995-e4fe-4355-89dc-52c4e3e13768" (UID: "c59f1995-e4fe-4355-89dc-52c4e3e13768"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.301993 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.302030 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vll4c\" (UniqueName: \"kubernetes.io/projected/c59f1995-e4fe-4355-89dc-52c4e3e13768-kube-api-access-vll4c\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.302046 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.302061 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59f1995-e4fe-4355-89dc-52c4e3e13768-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.302073 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c59f1995-e4fe-4355-89dc-52c4e3e13768-logs\") on node \"crc\" DevicePath \"\"" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.304997 4775 generic.go:334] "Generic (PLEG): container finished" podID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerID="5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b" exitCode=0 Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.305083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c59f1995-e4fe-4355-89dc-52c4e3e13768","Type":"ContainerDied","Data":"5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b"} Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.305121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c59f1995-e4fe-4355-89dc-52c4e3e13768","Type":"ContainerDied","Data":"88f299eae44d14624b299884f70903b778a2887a7f2d828f5de9fcf52140c815"} Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.305148 4775 scope.go:117] "RemoveContainer" containerID="5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.305304 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.316554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e8d44208-4c24-4835-89f6-a471ce50ddae","Type":"ContainerStarted","Data":"093ff6e93872303fdcec6415b91e68ae5b8f5e2e4f98d1050f2c385646471fb3"} Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.316584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e8d44208-4c24-4835-89f6-a471ce50ddae","Type":"ContainerStarted","Data":"7654588dab175a0903d1038819ffbd5b76517e597ab59cad1d7324d2d6532ddf"} Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.339242 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.339224171 podStartE2EDuration="2.339224171s" podCreationTimestamp="2025-11-26 06:44:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:29.334440744 +0000 UTC m=+1152.695744686" watchObservedRunningTime="2025-11-26 06:44:29.339224171 +0000 UTC m=+1152.700528123" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.351898 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ab86ff-2e8a-424f-9c94-6522b7e6ab00" path="/var/lib/kubelet/pods/c8ab86ff-2e8a-424f-9c94-6522b7e6ab00/volumes" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.353142 4775 scope.go:117] "RemoveContainer" containerID="279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.367793 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.380295 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.403784 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:29 crc kubenswrapper[4775]: E1126 06:44:29.406225 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-metadata" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.406337 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-metadata" Nov 26 06:44:29 crc kubenswrapper[4775]: E1126 06:44:29.406440 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-log" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.406549 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-log" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.406902 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-metadata" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.407005 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" containerName="nova-metadata-log" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.408303 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.410998 4775 scope.go:117] "RemoveContainer" containerID="5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.413276 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.413638 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 26 06:44:29 crc kubenswrapper[4775]: E1126 06:44:29.433945 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b\": container with ID starting with 5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b not found: ID does not exist" containerID="5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.434142 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b"} err="failed to get container status \"5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b\": rpc error: code = NotFound desc = could not find container \"5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b\": container with ID starting with 5ff1f7cbe62ae5136ee30432466c1208c14bd559442179479ab7a0de2de7cf9b not found: ID does not exist" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.434251 4775 scope.go:117] "RemoveContainer" containerID="279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de" Nov 26 06:44:29 crc kubenswrapper[4775]: E1126 06:44:29.434755 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de\": container with ID starting with 279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de not found: ID does not exist" containerID="279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.434850 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de"} err="failed to get container status \"279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de\": rpc error: code = NotFound desc = could not find container \"279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de\": container with ID starting with 279bd9a9027388a14e2f364810368aae52b2eb1ccebd8760921567387e7551de not found: ID does not exist" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.435856 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.506366 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.506651 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.506803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg7z2\" (UniqueName: \"kubernetes.io/projected/2d57ebfc-a725-46a3-9224-f011eadfc783-kube-api-access-dg7z2\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.506901 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d57ebfc-a725-46a3-9224-f011eadfc783-logs\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.507014 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-config-data\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.609326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-config-data\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.609436 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.609466 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.609532 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg7z2\" (UniqueName: \"kubernetes.io/projected/2d57ebfc-a725-46a3-9224-f011eadfc783-kube-api-access-dg7z2\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.609571 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d57ebfc-a725-46a3-9224-f011eadfc783-logs\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.610035 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d57ebfc-a725-46a3-9224-f011eadfc783-logs\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.613991 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-config-data\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.615064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.619462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d57ebfc-a725-46a3-9224-f011eadfc783-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.628480 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg7z2\" (UniqueName: \"kubernetes.io/projected/2d57ebfc-a725-46a3-9224-f011eadfc783-kube-api-access-dg7z2\") pod \"nova-metadata-0\" (UID: \"2d57ebfc-a725-46a3-9224-f011eadfc783\") " pod="openstack/nova-metadata-0" Nov 26 06:44:29 crc kubenswrapper[4775]: I1126 06:44:29.735641 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 26 06:44:30 crc kubenswrapper[4775]: I1126 06:44:30.248357 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 26 06:44:30 crc kubenswrapper[4775]: W1126 06:44:30.273228 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d57ebfc_a725_46a3_9224_f011eadfc783.slice/crio-070be48e0d3f1505e28644eda7632016dfc83aa15b6c0a96b58be26b110b72e5 WatchSource:0}: Error finding container 070be48e0d3f1505e28644eda7632016dfc83aa15b6c0a96b58be26b110b72e5: Status 404 returned error can't find the container with id 070be48e0d3f1505e28644eda7632016dfc83aa15b6c0a96b58be26b110b72e5 Nov 26 06:44:30 crc kubenswrapper[4775]: I1126 06:44:30.337968 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d57ebfc-a725-46a3-9224-f011eadfc783","Type":"ContainerStarted","Data":"070be48e0d3f1505e28644eda7632016dfc83aa15b6c0a96b58be26b110b72e5"} Nov 26 06:44:31 crc kubenswrapper[4775]: I1126 06:44:31.337159 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c59f1995-e4fe-4355-89dc-52c4e3e13768" path="/var/lib/kubelet/pods/c59f1995-e4fe-4355-89dc-52c4e3e13768/volumes" Nov 26 06:44:31 crc kubenswrapper[4775]: I1126 06:44:31.346384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d57ebfc-a725-46a3-9224-f011eadfc783","Type":"ContainerStarted","Data":"3af8c70ed661e19e61eda8e8e0b1020ce3c4bbee1bdc0856cdd3491fe99f1ec8"} Nov 26 06:44:31 crc kubenswrapper[4775]: I1126 06:44:31.346426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2d57ebfc-a725-46a3-9224-f011eadfc783","Type":"ContainerStarted","Data":"8a75bd12d0ac022186ade348f10ec86ad44cc47c1c0f2ef68e18445cc552bcc0"} Nov 26 06:44:31 crc kubenswrapper[4775]: I1126 06:44:31.377535 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.377492244 podStartE2EDuration="2.377492244s" podCreationTimestamp="2025-11-26 06:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:44:31.366942625 +0000 UTC m=+1154.728246597" watchObservedRunningTime="2025-11-26 06:44:31.377492244 +0000 UTC m=+1154.738796226" Nov 26 06:44:33 crc kubenswrapper[4775]: I1126 06:44:33.052015 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 26 06:44:34 crc kubenswrapper[4775]: I1126 06:44:34.736770 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 06:44:34 crc kubenswrapper[4775]: I1126 06:44:34.737129 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 26 06:44:36 crc kubenswrapper[4775]: I1126 06:44:36.682151 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 06:44:36 crc kubenswrapper[4775]: I1126 06:44:36.682198 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 26 06:44:37 crc kubenswrapper[4775]: I1126 06:44:37.696007 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="09a96730-d166-4111-b88a-b2a51ca06b5a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:37 crc kubenswrapper[4775]: I1126 06:44:37.696354 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="09a96730-d166-4111-b88a-b2a51ca06b5a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:38 crc kubenswrapper[4775]: I1126 06:44:38.051513 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 26 06:44:38 crc kubenswrapper[4775]: I1126 06:44:38.080638 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 26 06:44:38 crc kubenswrapper[4775]: I1126 06:44:38.454632 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 26 06:44:39 crc kubenswrapper[4775]: I1126 06:44:39.737011 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 06:44:39 crc kubenswrapper[4775]: I1126 06:44:39.737440 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 26 06:44:40 crc kubenswrapper[4775]: I1126 06:44:40.752932 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2d57ebfc-a725-46a3-9224-f011eadfc783" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:40 crc kubenswrapper[4775]: I1126 06:44:40.753037 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2d57ebfc-a725-46a3-9224-f011eadfc783" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 26 06:44:44 crc kubenswrapper[4775]: I1126 06:44:44.514899 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 26 06:44:46 crc kubenswrapper[4775]: I1126 06:44:46.697701 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 06:44:46 crc kubenswrapper[4775]: I1126 06:44:46.698394 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 06:44:46 crc kubenswrapper[4775]: I1126 06:44:46.710192 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 06:44:46 crc kubenswrapper[4775]: I1126 06:44:46.711076 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 26 06:44:47 crc kubenswrapper[4775]: I1126 06:44:47.539059 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 26 06:44:47 crc kubenswrapper[4775]: I1126 06:44:47.563370 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 26 06:44:49 crc kubenswrapper[4775]: I1126 06:44:49.742292 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 26 06:44:49 crc kubenswrapper[4775]: I1126 06:44:49.744274 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 26 06:44:49 crc kubenswrapper[4775]: I1126 06:44:49.748908 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 26 06:44:50 crc kubenswrapper[4775]: I1126 06:44:50.566463 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 26 06:44:58 crc kubenswrapper[4775]: I1126 06:44:58.599051 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:44:59 crc kubenswrapper[4775]: I1126 06:44:59.499861 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.160611 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9"] Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.162387 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.164302 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.164818 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.173265 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9"] Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.256126 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86378490-4d35-4d6f-8b0e-16225f624014-secret-volume\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.256183 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hpsd\" (UniqueName: \"kubernetes.io/projected/86378490-4d35-4d6f-8b0e-16225f624014-kube-api-access-9hpsd\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.256323 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86378490-4d35-4d6f-8b0e-16225f624014-config-volume\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.357833 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86378490-4d35-4d6f-8b0e-16225f624014-secret-volume\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.357883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hpsd\" (UniqueName: \"kubernetes.io/projected/86378490-4d35-4d6f-8b0e-16225f624014-kube-api-access-9hpsd\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.357966 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86378490-4d35-4d6f-8b0e-16225f624014-config-volume\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.358917 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86378490-4d35-4d6f-8b0e-16225f624014-config-volume\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.368075 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86378490-4d35-4d6f-8b0e-16225f624014-secret-volume\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.384499 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hpsd\" (UniqueName: \"kubernetes.io/projected/86378490-4d35-4d6f-8b0e-16225f624014-kube-api-access-9hpsd\") pod \"collect-profiles-29402325-mkjf9\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.484567 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:00 crc kubenswrapper[4775]: I1126 06:45:00.982408 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9"] Nov 26 06:45:01 crc kubenswrapper[4775]: I1126 06:45:01.668229 4775 generic.go:334] "Generic (PLEG): container finished" podID="86378490-4d35-4d6f-8b0e-16225f624014" containerID="19833754e9909e7e7724dd83c173f935a074c4d511545d0452cfbb02fa8b3835" exitCode=0 Nov 26 06:45:01 crc kubenswrapper[4775]: I1126 06:45:01.668502 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" event={"ID":"86378490-4d35-4d6f-8b0e-16225f624014","Type":"ContainerDied","Data":"19833754e9909e7e7724dd83c173f935a074c4d511545d0452cfbb02fa8b3835"} Nov 26 06:45:01 crc kubenswrapper[4775]: I1126 06:45:01.668529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" event={"ID":"86378490-4d35-4d6f-8b0e-16225f624014","Type":"ContainerStarted","Data":"c14f943055ed84b9243afba097c81023e3fe375650659edad2c653e1f129a417"} Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.024401 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="rabbitmq" containerID="cri-o://ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804" gracePeriod=604796 Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.067834 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.215436 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hpsd\" (UniqueName: \"kubernetes.io/projected/86378490-4d35-4d6f-8b0e-16225f624014-kube-api-access-9hpsd\") pod \"86378490-4d35-4d6f-8b0e-16225f624014\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.215770 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86378490-4d35-4d6f-8b0e-16225f624014-secret-volume\") pod \"86378490-4d35-4d6f-8b0e-16225f624014\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.215810 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86378490-4d35-4d6f-8b0e-16225f624014-config-volume\") pod \"86378490-4d35-4d6f-8b0e-16225f624014\" (UID: \"86378490-4d35-4d6f-8b0e-16225f624014\") " Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.216426 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86378490-4d35-4d6f-8b0e-16225f624014-config-volume" (OuterVolumeSpecName: "config-volume") pod "86378490-4d35-4d6f-8b0e-16225f624014" (UID: "86378490-4d35-4d6f-8b0e-16225f624014"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.223945 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86378490-4d35-4d6f-8b0e-16225f624014-kube-api-access-9hpsd" (OuterVolumeSpecName: "kube-api-access-9hpsd") pod "86378490-4d35-4d6f-8b0e-16225f624014" (UID: "86378490-4d35-4d6f-8b0e-16225f624014"). InnerVolumeSpecName "kube-api-access-9hpsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.224448 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86378490-4d35-4d6f-8b0e-16225f624014-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "86378490-4d35-4d6f-8b0e-16225f624014" (UID: "86378490-4d35-4d6f-8b0e-16225f624014"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.317568 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hpsd\" (UniqueName: \"kubernetes.io/projected/86378490-4d35-4d6f-8b0e-16225f624014-kube-api-access-9hpsd\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.317604 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86378490-4d35-4d6f-8b0e-16225f624014-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.317613 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86378490-4d35-4d6f-8b0e-16225f624014-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:03 crc kubenswrapper[4775]: I1126 06:45:03.667372 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Nov 26 06:45:04 crc kubenswrapper[4775]: I1126 06:45:04.124177 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" event={"ID":"86378490-4d35-4d6f-8b0e-16225f624014","Type":"ContainerDied","Data":"c14f943055ed84b9243afba097c81023e3fe375650659edad2c653e1f129a417"} Nov 26 06:45:04 crc kubenswrapper[4775]: I1126 06:45:04.124224 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c14f943055ed84b9243afba097c81023e3fe375650659edad2c653e1f129a417" Nov 26 06:45:04 crc kubenswrapper[4775]: I1126 06:45:04.124284 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9" Nov 26 06:45:04 crc kubenswrapper[4775]: I1126 06:45:04.697489 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerName="rabbitmq" containerID="cri-o://23f1b68e58af700cd8d63e29753ae5ef210dc630aedd23de681e823f78054454" gracePeriod=604795 Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.765223 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856037 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr5v9\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-kube-api-access-jr5v9\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856085 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-tls\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856118 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-confd\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856147 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63f863dc-36a4-4325-b520-e26b3dca309e-erlang-cookie-secret\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856212 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63f863dc-36a4-4325-b520-e26b3dca309e-pod-info\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856233 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-plugins-conf\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856288 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-plugins\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-server-conf\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856381 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-config-data\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.856431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-erlang-cookie\") pod \"63f863dc-36a4-4325-b520-e26b3dca309e\" (UID: \"63f863dc-36a4-4325-b520-e26b3dca309e\") " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.857273 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.859531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.859852 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.864135 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-kube-api-access-jr5v9" (OuterVolumeSpecName: "kube-api-access-jr5v9") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "kube-api-access-jr5v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.864602 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.867053 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.867885 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/63f863dc-36a4-4325-b520-e26b3dca309e-pod-info" (OuterVolumeSpecName: "pod-info") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.870330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63f863dc-36a4-4325-b520-e26b3dca309e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.900486 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-config-data" (OuterVolumeSpecName: "config-data") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.941328 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-server-conf" (OuterVolumeSpecName: "server-conf") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958601 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr5v9\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-kube-api-access-jr5v9\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958641 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958656 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/63f863dc-36a4-4325-b520-e26b3dca309e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958689 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958700 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/63f863dc-36a4-4325-b520-e26b3dca309e-pod-info\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958729 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958740 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958752 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-server-conf\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958762 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/63f863dc-36a4-4325-b520-e26b3dca309e-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.958771 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.989372 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 26 06:45:09 crc kubenswrapper[4775]: I1126 06:45:09.998004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "63f863dc-36a4-4325-b520-e26b3dca309e" (UID: "63f863dc-36a4-4325-b520-e26b3dca309e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.061386 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/63f863dc-36a4-4325-b520-e26b3dca309e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.061416 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.192874 4775 generic.go:334] "Generic (PLEG): container finished" podID="63f863dc-36a4-4325-b520-e26b3dca309e" containerID="ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804" exitCode=0 Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.192926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63f863dc-36a4-4325-b520-e26b3dca309e","Type":"ContainerDied","Data":"ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804"} Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.192959 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"63f863dc-36a4-4325-b520-e26b3dca309e","Type":"ContainerDied","Data":"f5332c832f6e8c06b03d96af0eb0b1a552b4cd1045106f0c2f4cdb19e13ff2db"} Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.192981 4775 scope.go:117] "RemoveContainer" containerID="ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.192994 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.222124 4775 scope.go:117] "RemoveContainer" containerID="986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.238541 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.243784 4775 scope.go:117] "RemoveContainer" containerID="ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804" Nov 26 06:45:10 crc kubenswrapper[4775]: E1126 06:45:10.244298 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804\": container with ID starting with ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804 not found: ID does not exist" containerID="ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.244338 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804"} err="failed to get container status \"ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804\": rpc error: code = NotFound desc = could not find container \"ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804\": container with ID starting with ae7ce733fab89aabe6e274047ced26c90764e81eb0c932e2e7e163ff1e4e6804 not found: ID does not exist" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.244366 4775 scope.go:117] "RemoveContainer" containerID="986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf" Nov 26 06:45:10 crc kubenswrapper[4775]: E1126 06:45:10.244819 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf\": container with ID starting with 986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf not found: ID does not exist" containerID="986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.244853 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf"} err="failed to get container status \"986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf\": rpc error: code = NotFound desc = could not find container \"986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf\": container with ID starting with 986b8f97cb71b88e6221a392ad4278209c5485d438019f30afd60d51751fd9cf not found: ID does not exist" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.263677 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.278205 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:45:10 crc kubenswrapper[4775]: E1126 06:45:10.278649 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="rabbitmq" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.278667 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="rabbitmq" Nov 26 06:45:10 crc kubenswrapper[4775]: E1126 06:45:10.278694 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="setup-container" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.278703 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="setup-container" Nov 26 06:45:10 crc kubenswrapper[4775]: E1126 06:45:10.278736 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86378490-4d35-4d6f-8b0e-16225f624014" containerName="collect-profiles" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.278744 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="86378490-4d35-4d6f-8b0e-16225f624014" containerName="collect-profiles" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.278911 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" containerName="rabbitmq" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.278933 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="86378490-4d35-4d6f-8b0e-16225f624014" containerName="collect-profiles" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.280128 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.283537 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.283743 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.283937 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.284109 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.284239 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.284347 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.284457 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xrgm4" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.292978 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.366695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.366844 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/837ab270-d963-4406-9a55-390bf0611e14-pod-info\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.366891 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.366928 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.366965 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwwjx\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-kube-api-access-kwwjx\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.366991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.367181 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-config-data\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.367307 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/837ab270-d963-4406-9a55-390bf0611e14-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.367354 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.367391 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-server-conf\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.367481 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.468926 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469026 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwwjx\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-kube-api-access-kwwjx\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469064 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-config-data\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469542 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/837ab270-d963-4406-9a55-390bf0611e14-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469573 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469596 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-server-conf\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469809 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469864 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.469879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/837ab270-d963-4406-9a55-390bf0611e14-pod-info\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.470017 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.470078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.470480 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-config-data\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.470798 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.471445 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.471902 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/837ab270-d963-4406-9a55-390bf0611e14-server-conf\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.474290 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.475089 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.477206 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/837ab270-d963-4406-9a55-390bf0611e14-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.484199 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/837ab270-d963-4406-9a55-390bf0611e14-pod-info\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.491821 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwwjx\" (UniqueName: \"kubernetes.io/projected/837ab270-d963-4406-9a55-390bf0611e14-kube-api-access-kwwjx\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.499302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"837ab270-d963-4406-9a55-390bf0611e14\") " pod="openstack/rabbitmq-server-0" Nov 26 06:45:10 crc kubenswrapper[4775]: I1126 06:45:10.646288 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.217969 4775 generic.go:334] "Generic (PLEG): container finished" podID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerID="23f1b68e58af700cd8d63e29753ae5ef210dc630aedd23de681e823f78054454" exitCode=0 Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.218254 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"92a5651e-7175-45ca-a69b-dc19a11f2943","Type":"ContainerDied","Data":"23f1b68e58af700cd8d63e29753ae5ef210dc630aedd23de681e823f78054454"} Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.251451 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.343601 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63f863dc-36a4-4325-b520-e26b3dca309e" path="/var/lib/kubelet/pods/63f863dc-36a4-4325-b520-e26b3dca309e/volumes" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.481293 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.598756 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-confd\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.598837 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92a5651e-7175-45ca-a69b-dc19a11f2943-erlang-cookie-secret\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.598897 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92a5651e-7175-45ca-a69b-dc19a11f2943-pod-info\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.598924 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-config-data\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599041 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-server-conf\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599086 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-plugins\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-979q2\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-kube-api-access-979q2\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-plugins-conf\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-tls\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.599257 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-erlang-cookie\") pod \"92a5651e-7175-45ca-a69b-dc19a11f2943\" (UID: \"92a5651e-7175-45ca-a69b-dc19a11f2943\") " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.600319 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.603945 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/92a5651e-7175-45ca-a69b-dc19a11f2943-pod-info" (OuterVolumeSpecName: "pod-info") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.604290 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.605456 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.607524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.612576 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.612642 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-kube-api-access-979q2" (OuterVolumeSpecName: "kube-api-access-979q2") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "kube-api-access-979q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.622838 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a5651e-7175-45ca-a69b-dc19a11f2943-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.675439 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-config-data" (OuterVolumeSpecName: "config-data") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.679760 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-server-conf" (OuterVolumeSpecName: "server-conf") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701610 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-979q2\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-kube-api-access-979q2\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701646 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701656 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701665 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701674 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/92a5651e-7175-45ca-a69b-dc19a11f2943-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701683 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/92a5651e-7175-45ca-a69b-dc19a11f2943-pod-info\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701691 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701729 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701739 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/92a5651e-7175-45ca-a69b-dc19a11f2943-server-conf\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.701747 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.732988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "92a5651e-7175-45ca-a69b-dc19a11f2943" (UID: "92a5651e-7175-45ca-a69b-dc19a11f2943"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.733656 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.803241 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/92a5651e-7175-45ca-a69b-dc19a11f2943-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.803271 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.905866 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-2tz2k"] Nov 26 06:45:11 crc kubenswrapper[4775]: E1126 06:45:11.906242 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerName="setup-container" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.906257 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerName="setup-container" Nov 26 06:45:11 crc kubenswrapper[4775]: E1126 06:45:11.906284 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerName="rabbitmq" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.906291 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerName="rabbitmq" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.906486 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" containerName="rabbitmq" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.907520 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.919386 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-2tz2k"] Nov 26 06:45:11 crc kubenswrapper[4775]: I1126 06:45:11.944343 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006400 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-svc\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006449 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-config\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006486 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqbf7\" (UniqueName: \"kubernetes.io/projected/5353ea9f-0435-4ea6-b150-175533a0792f-kube-api-access-dqbf7\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006580 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006605 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006659 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.006675 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.108068 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109488 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109651 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-svc\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-config\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqbf7\" (UniqueName: \"kubernetes.io/projected/5353ea9f-0435-4ea6-b150-175533a0792f-kube-api-access-dqbf7\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.110747 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.109424 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.111411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.112126 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-svc\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.112807 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.113052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-config\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.144291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqbf7\" (UniqueName: \"kubernetes.io/projected/5353ea9f-0435-4ea6-b150-175533a0792f-kube-api-access-dqbf7\") pod \"dnsmasq-dns-67b789f86c-2tz2k\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.238143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"837ab270-d963-4406-9a55-390bf0611e14","Type":"ContainerStarted","Data":"8fd4d747a9b43d4d6757ace52bc3684a149f536c51777f0b2b682e6660f6f3e8"} Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.241176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"92a5651e-7175-45ca-a69b-dc19a11f2943","Type":"ContainerDied","Data":"501a9d556dc52a76e2cc46f1734a0285e42c7fcf5e57d92b73892b6703aa0a95"} Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.241216 4775 scope.go:117] "RemoveContainer" containerID="23f1b68e58af700cd8d63e29753ae5ef210dc630aedd23de681e823f78054454" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.241363 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.268056 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.289482 4775 scope.go:117] "RemoveContainer" containerID="bcdd2a68e5172c990298b42de9bed69bd8c00801bcd6c8e980fd5d7cc6fafff6" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.299135 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.322078 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.344685 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.346391 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.349142 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.349342 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.352051 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qc8lq" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.352352 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.352523 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.352664 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.353026 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.354435 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422498 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422618 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422661 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422696 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvrhw\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-kube-api-access-zvrhw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.422843 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524695 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524791 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524860 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524957 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.524993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.525042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.525067 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.525089 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvrhw\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-kube-api-access-zvrhw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.525136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.526076 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.526986 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.531107 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.534510 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.535315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.535339 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.562409 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.562537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.563774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.564407 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.566126 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvrhw\" (UniqueName: \"kubernetes.io/projected/ce90c242-2ac3-4509-bf08-2f7c77b1aff0-kube-api-access-zvrhw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.604658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ce90c242-2ac3-4509-bf08-2f7c77b1aff0\") " pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.689896 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:12 crc kubenswrapper[4775]: I1126 06:45:12.828923 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-2tz2k"] Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.171750 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.278191 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce90c242-2ac3-4509-bf08-2f7c77b1aff0","Type":"ContainerStarted","Data":"b10dbe3651e0265a009fbe9a958679ff2fcbe445d6cf7ff6f9e80b9f5c016d96"} Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.296971 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"837ab270-d963-4406-9a55-390bf0611e14","Type":"ContainerStarted","Data":"ad7d2128db1a5aed8e70c96d71abcc87e04db791581fd0cf15c2f55681712b9b"} Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.311530 4775 generic.go:334] "Generic (PLEG): container finished" podID="5353ea9f-0435-4ea6-b150-175533a0792f" containerID="5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1" exitCode=0 Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.311584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" event={"ID":"5353ea9f-0435-4ea6-b150-175533a0792f","Type":"ContainerDied","Data":"5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1"} Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.311617 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" event={"ID":"5353ea9f-0435-4ea6-b150-175533a0792f","Type":"ContainerStarted","Data":"f8a11539bc445e56ce94158c4157f42457931ae0663f73fcdddd93f79153714c"} Nov 26 06:45:13 crc kubenswrapper[4775]: I1126 06:45:13.388000 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a5651e-7175-45ca-a69b-dc19a11f2943" path="/var/lib/kubelet/pods/92a5651e-7175-45ca-a69b-dc19a11f2943/volumes" Nov 26 06:45:14 crc kubenswrapper[4775]: I1126 06:45:14.331010 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" event={"ID":"5353ea9f-0435-4ea6-b150-175533a0792f","Type":"ContainerStarted","Data":"2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77"} Nov 26 06:45:14 crc kubenswrapper[4775]: I1126 06:45:14.363436 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" podStartSLOduration=3.363417755 podStartE2EDuration="3.363417755s" podCreationTimestamp="2025-11-26 06:45:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:45:14.36014757 +0000 UTC m=+1197.721451592" watchObservedRunningTime="2025-11-26 06:45:14.363417755 +0000 UTC m=+1197.724721707" Nov 26 06:45:15 crc kubenswrapper[4775]: I1126 06:45:15.363074 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:15 crc kubenswrapper[4775]: I1126 06:45:15.363148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce90c242-2ac3-4509-bf08-2f7c77b1aff0","Type":"ContainerStarted","Data":"5648a267f6824e1623904f5a0feda23ba40a11d090523983164a3e20af32cc17"} Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.269959 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.338042 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-qkdkn"] Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.338361 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" podUID="fcee54b8-a240-4821-82da-40e6649800d9" containerName="dnsmasq-dns" containerID="cri-o://0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962" gracePeriod=10 Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.498071 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-dhkrg"] Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.500070 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.521993 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-dhkrg"] Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.546938 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tztzh\" (UniqueName: \"kubernetes.io/projected/e837eeaf-6cf5-44e0-b183-334f62304bc7-kube-api-access-tztzh\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.546995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.547038 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.547101 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-config\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.547134 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.547173 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.547196 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.648997 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.649112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tztzh\" (UniqueName: \"kubernetes.io/projected/e837eeaf-6cf5-44e0-b183-334f62304bc7-kube-api-access-tztzh\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.649151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.649188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.649229 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-config\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.649276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.649335 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.650445 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.650846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.651069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.651538 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.651601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-config\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.652130 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e837eeaf-6cf5-44e0-b183-334f62304bc7-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.670290 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tztzh\" (UniqueName: \"kubernetes.io/projected/e837eeaf-6cf5-44e0-b183-334f62304bc7-kube-api-access-tztzh\") pod \"dnsmasq-dns-cb6ffcf87-dhkrg\" (UID: \"e837eeaf-6cf5-44e0-b183-334f62304bc7\") " pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.820015 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.920567 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.955573 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-sb\") pod \"fcee54b8-a240-4821-82da-40e6649800d9\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.955740 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvdkf\" (UniqueName: \"kubernetes.io/projected/fcee54b8-a240-4821-82da-40e6649800d9-kube-api-access-hvdkf\") pod \"fcee54b8-a240-4821-82da-40e6649800d9\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.955775 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-config\") pod \"fcee54b8-a240-4821-82da-40e6649800d9\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.955813 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-svc\") pod \"fcee54b8-a240-4821-82da-40e6649800d9\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.955844 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-swift-storage-0\") pod \"fcee54b8-a240-4821-82da-40e6649800d9\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.955967 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-nb\") pod \"fcee54b8-a240-4821-82da-40e6649800d9\" (UID: \"fcee54b8-a240-4821-82da-40e6649800d9\") " Nov 26 06:45:22 crc kubenswrapper[4775]: I1126 06:45:22.967482 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcee54b8-a240-4821-82da-40e6649800d9-kube-api-access-hvdkf" (OuterVolumeSpecName: "kube-api-access-hvdkf") pod "fcee54b8-a240-4821-82da-40e6649800d9" (UID: "fcee54b8-a240-4821-82da-40e6649800d9"). InnerVolumeSpecName "kube-api-access-hvdkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.007493 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fcee54b8-a240-4821-82da-40e6649800d9" (UID: "fcee54b8-a240-4821-82da-40e6649800d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.013770 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-config" (OuterVolumeSpecName: "config") pod "fcee54b8-a240-4821-82da-40e6649800d9" (UID: "fcee54b8-a240-4821-82da-40e6649800d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.023074 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fcee54b8-a240-4821-82da-40e6649800d9" (UID: "fcee54b8-a240-4821-82da-40e6649800d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.036858 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fcee54b8-a240-4821-82da-40e6649800d9" (UID: "fcee54b8-a240-4821-82da-40e6649800d9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.044474 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fcee54b8-a240-4821-82da-40e6649800d9" (UID: "fcee54b8-a240-4821-82da-40e6649800d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.058256 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.058285 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvdkf\" (UniqueName: \"kubernetes.io/projected/fcee54b8-a240-4821-82da-40e6649800d9-kube-api-access-hvdkf\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.058295 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.058304 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.058311 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.058320 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fcee54b8-a240-4821-82da-40e6649800d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.350308 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-dhkrg"] Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.445460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" event={"ID":"e837eeaf-6cf5-44e0-b183-334f62304bc7","Type":"ContainerStarted","Data":"329910e2813a1f4a0347fe1d28bfe85db364c9b25b195ead6f944439dd7792dd"} Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.448048 4775 generic.go:334] "Generic (PLEG): container finished" podID="fcee54b8-a240-4821-82da-40e6649800d9" containerID="0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962" exitCode=0 Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.448084 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" event={"ID":"fcee54b8-a240-4821-82da-40e6649800d9","Type":"ContainerDied","Data":"0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962"} Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.448103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" event={"ID":"fcee54b8-a240-4821-82da-40e6649800d9","Type":"ContainerDied","Data":"a08286bf6e21004210b12924cf89621db3aebc93bc8a2befa1bde5c3dce1f4fa"} Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.448122 4775 scope.go:117] "RemoveContainer" containerID="0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.448269 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-qkdkn" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.555502 4775 scope.go:117] "RemoveContainer" containerID="bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.558629 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-qkdkn"] Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.566332 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-qkdkn"] Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.601521 4775 scope.go:117] "RemoveContainer" containerID="0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962" Nov 26 06:45:23 crc kubenswrapper[4775]: E1126 06:45:23.602019 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962\": container with ID starting with 0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962 not found: ID does not exist" containerID="0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.602060 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962"} err="failed to get container status \"0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962\": rpc error: code = NotFound desc = could not find container \"0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962\": container with ID starting with 0892db51ba9d105b354d80dee40584d9cc7851ead93e3616540bdd7e8660b962 not found: ID does not exist" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.602084 4775 scope.go:117] "RemoveContainer" containerID="bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e" Nov 26 06:45:23 crc kubenswrapper[4775]: E1126 06:45:23.602436 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e\": container with ID starting with bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e not found: ID does not exist" containerID="bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e" Nov 26 06:45:23 crc kubenswrapper[4775]: I1126 06:45:23.602483 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e"} err="failed to get container status \"bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e\": rpc error: code = NotFound desc = could not find container \"bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e\": container with ID starting with bbfe523fb96ba86bd240ce788721e73b593d483a601fe20250e4367e66c4b91e not found: ID does not exist" Nov 26 06:45:24 crc kubenswrapper[4775]: I1126 06:45:24.467009 4775 generic.go:334] "Generic (PLEG): container finished" podID="e837eeaf-6cf5-44e0-b183-334f62304bc7" containerID="09afd3c8fbcbc0694f6dec8e3d31ea91bed9f86bd6eee412b3faab8ef7dffb18" exitCode=0 Nov 26 06:45:24 crc kubenswrapper[4775]: I1126 06:45:24.467092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" event={"ID":"e837eeaf-6cf5-44e0-b183-334f62304bc7","Type":"ContainerDied","Data":"09afd3c8fbcbc0694f6dec8e3d31ea91bed9f86bd6eee412b3faab8ef7dffb18"} Nov 26 06:45:25 crc kubenswrapper[4775]: I1126 06:45:25.363151 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcee54b8-a240-4821-82da-40e6649800d9" path="/var/lib/kubelet/pods/fcee54b8-a240-4821-82da-40e6649800d9/volumes" Nov 26 06:45:25 crc kubenswrapper[4775]: I1126 06:45:25.486877 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" event={"ID":"e837eeaf-6cf5-44e0-b183-334f62304bc7","Type":"ContainerStarted","Data":"5ceee885e04082c3df4f3968348957491952c39177ee7fabc4d77b4741bfc389"} Nov 26 06:45:25 crc kubenswrapper[4775]: I1126 06:45:25.487227 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:25 crc kubenswrapper[4775]: I1126 06:45:25.522627 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" podStartSLOduration=3.522601641 podStartE2EDuration="3.522601641s" podCreationTimestamp="2025-11-26 06:45:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:45:25.50843073 +0000 UTC m=+1208.869734722" watchObservedRunningTime="2025-11-26 06:45:25.522601641 +0000 UTC m=+1208.883905623" Nov 26 06:45:32 crc kubenswrapper[4775]: I1126 06:45:32.822176 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-dhkrg" Nov 26 06:45:32 crc kubenswrapper[4775]: I1126 06:45:32.927547 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-2tz2k"] Nov 26 06:45:32 crc kubenswrapper[4775]: I1126 06:45:32.935139 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" containerName="dnsmasq-dns" containerID="cri-o://2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77" gracePeriod=10 Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.436691 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.534601 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-config\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.534743 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqbf7\" (UniqueName: \"kubernetes.io/projected/5353ea9f-0435-4ea6-b150-175533a0792f-kube-api-access-dqbf7\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.534779 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-sb\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.534959 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-nb\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.535099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-svc\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.535178 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-swift-storage-0\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.535260 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-openstack-edpm-ipam\") pod \"5353ea9f-0435-4ea6-b150-175533a0792f\" (UID: \"5353ea9f-0435-4ea6-b150-175533a0792f\") " Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.541850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5353ea9f-0435-4ea6-b150-175533a0792f-kube-api-access-dqbf7" (OuterVolumeSpecName: "kube-api-access-dqbf7") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "kube-api-access-dqbf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.598196 4775 generic.go:334] "Generic (PLEG): container finished" podID="5353ea9f-0435-4ea6-b150-175533a0792f" containerID="2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77" exitCode=0 Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.598267 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" event={"ID":"5353ea9f-0435-4ea6-b150-175533a0792f","Type":"ContainerDied","Data":"2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77"} Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.598309 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" event={"ID":"5353ea9f-0435-4ea6-b150-175533a0792f","Type":"ContainerDied","Data":"f8a11539bc445e56ce94158c4157f42457931ae0663f73fcdddd93f79153714c"} Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.598355 4775 scope.go:117] "RemoveContainer" containerID="2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.598628 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-2tz2k" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.612740 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.620872 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-config" (OuterVolumeSpecName: "config") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.625108 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.635317 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637291 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637652 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637685 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-config\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637697 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqbf7\" (UniqueName: \"kubernetes.io/projected/5353ea9f-0435-4ea6-b150-175533a0792f-kube-api-access-dqbf7\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637707 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637772 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.637781 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.639177 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5353ea9f-0435-4ea6-b150-175533a0792f" (UID: "5353ea9f-0435-4ea6-b150-175533a0792f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.682396 4775 scope.go:117] "RemoveContainer" containerID="5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.704025 4775 scope.go:117] "RemoveContainer" containerID="2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77" Nov 26 06:45:33 crc kubenswrapper[4775]: E1126 06:45:33.704467 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77\": container with ID starting with 2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77 not found: ID does not exist" containerID="2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.704524 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77"} err="failed to get container status \"2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77\": rpc error: code = NotFound desc = could not find container \"2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77\": container with ID starting with 2968d12def69b1dde209aa5348dc6d672e4355ea6ae02f64e507b0cd9774be77 not found: ID does not exist" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.704569 4775 scope.go:117] "RemoveContainer" containerID="5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1" Nov 26 06:45:33 crc kubenswrapper[4775]: E1126 06:45:33.704994 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1\": container with ID starting with 5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1 not found: ID does not exist" containerID="5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.705036 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1"} err="failed to get container status \"5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1\": rpc error: code = NotFound desc = could not find container \"5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1\": container with ID starting with 5d2222adac6029479e814d22a69ad391c686ece49166c8e4f0c2c409335b10c1 not found: ID does not exist" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.739861 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5353ea9f-0435-4ea6-b150-175533a0792f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.937292 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-2tz2k"] Nov 26 06:45:33 crc kubenswrapper[4775]: I1126 06:45:33.946324 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-2tz2k"] Nov 26 06:45:35 crc kubenswrapper[4775]: I1126 06:45:35.351093 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" path="/var/lib/kubelet/pods/5353ea9f-0435-4ea6-b150-175533a0792f/volumes" Nov 26 06:45:45 crc kubenswrapper[4775]: I1126 06:45:45.736512 4775 generic.go:334] "Generic (PLEG): container finished" podID="837ab270-d963-4406-9a55-390bf0611e14" containerID="ad7d2128db1a5aed8e70c96d71abcc87e04db791581fd0cf15c2f55681712b9b" exitCode=0 Nov 26 06:45:45 crc kubenswrapper[4775]: I1126 06:45:45.736614 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"837ab270-d963-4406-9a55-390bf0611e14","Type":"ContainerDied","Data":"ad7d2128db1a5aed8e70c96d71abcc87e04db791581fd0cf15c2f55681712b9b"} Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.304772 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57"] Nov 26 06:45:46 crc kubenswrapper[4775]: E1126 06:45:46.305400 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcee54b8-a240-4821-82da-40e6649800d9" containerName="dnsmasq-dns" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.305414 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcee54b8-a240-4821-82da-40e6649800d9" containerName="dnsmasq-dns" Nov 26 06:45:46 crc kubenswrapper[4775]: E1126 06:45:46.305428 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" containerName="init" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.305436 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" containerName="init" Nov 26 06:45:46 crc kubenswrapper[4775]: E1126 06:45:46.305468 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" containerName="dnsmasq-dns" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.305477 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" containerName="dnsmasq-dns" Nov 26 06:45:46 crc kubenswrapper[4775]: E1126 06:45:46.305493 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcee54b8-a240-4821-82da-40e6649800d9" containerName="init" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.305503 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcee54b8-a240-4821-82da-40e6649800d9" containerName="init" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.305784 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcee54b8-a240-4821-82da-40e6649800d9" containerName="dnsmasq-dns" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.305807 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5353ea9f-0435-4ea6-b150-175533a0792f" containerName="dnsmasq-dns" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.306495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.308634 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.309196 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.309613 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.309956 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.329181 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57"] Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.403945 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.404044 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.404072 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.404259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt7nc\" (UniqueName: \"kubernetes.io/projected/561d8a6c-3d90-4220-b6ac-386d81f5016f-kube-api-access-gt7nc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.505798 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt7nc\" (UniqueName: \"kubernetes.io/projected/561d8a6c-3d90-4220-b6ac-386d81f5016f-kube-api-access-gt7nc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.505953 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.506043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.506082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.512033 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.513395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.513427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.528100 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt7nc\" (UniqueName: \"kubernetes.io/projected/561d8a6c-3d90-4220-b6ac-386d81f5016f-kube-api-access-gt7nc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.626607 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.754037 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"837ab270-d963-4406-9a55-390bf0611e14","Type":"ContainerStarted","Data":"7608894ae607510fb0aaeb2a87ba57bdc9fb47d93767c80b158cd38879a8d449"} Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.754686 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 26 06:45:46 crc kubenswrapper[4775]: I1126 06:45:46.806521 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.806503417 podStartE2EDuration="36.806503417s" podCreationTimestamp="2025-11-26 06:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:45:46.788036884 +0000 UTC m=+1230.149340846" watchObservedRunningTime="2025-11-26 06:45:46.806503417 +0000 UTC m=+1230.167807369" Nov 26 06:45:47 crc kubenswrapper[4775]: I1126 06:45:47.178365 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57"] Nov 26 06:45:47 crc kubenswrapper[4775]: I1126 06:45:47.191870 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 06:45:47 crc kubenswrapper[4775]: I1126 06:45:47.775982 4775 generic.go:334] "Generic (PLEG): container finished" podID="ce90c242-2ac3-4509-bf08-2f7c77b1aff0" containerID="5648a267f6824e1623904f5a0feda23ba40a11d090523983164a3e20af32cc17" exitCode=0 Nov 26 06:45:47 crc kubenswrapper[4775]: I1126 06:45:47.776084 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce90c242-2ac3-4509-bf08-2f7c77b1aff0","Type":"ContainerDied","Data":"5648a267f6824e1623904f5a0feda23ba40a11d090523983164a3e20af32cc17"} Nov 26 06:45:47 crc kubenswrapper[4775]: I1126 06:45:47.779123 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" event={"ID":"561d8a6c-3d90-4220-b6ac-386d81f5016f","Type":"ContainerStarted","Data":"dcff3387ad051119e3ab19438b4c2c6c66c0e906e82c87953af376af3ac9c470"} Nov 26 06:45:48 crc kubenswrapper[4775]: I1126 06:45:48.798273 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ce90c242-2ac3-4509-bf08-2f7c77b1aff0","Type":"ContainerStarted","Data":"14b1a98e04099e107b2c2bdfef8cce6f6e76500eed89fe752ea5c51835e41bb0"} Nov 26 06:45:48 crc kubenswrapper[4775]: I1126 06:45:48.798995 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:45:48 crc kubenswrapper[4775]: I1126 06:45:48.826210 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.82618814 podStartE2EDuration="36.82618814s" podCreationTimestamp="2025-11-26 06:45:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 06:45:48.824845695 +0000 UTC m=+1232.186149867" watchObservedRunningTime="2025-11-26 06:45:48.82618814 +0000 UTC m=+1232.187492102" Nov 26 06:45:51 crc kubenswrapper[4775]: I1126 06:45:51.419342 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:45:51 crc kubenswrapper[4775]: I1126 06:45:51.419993 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:45:58 crc kubenswrapper[4775]: I1126 06:45:58.912322 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" event={"ID":"561d8a6c-3d90-4220-b6ac-386d81f5016f","Type":"ContainerStarted","Data":"88e3d34cd44a362361fab1ffa9bb55208689d59183488afd0a644b02143b8eb7"} Nov 26 06:45:58 crc kubenswrapper[4775]: I1126 06:45:58.935773 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" podStartSLOduration=2.157659136 podStartE2EDuration="12.935755783s" podCreationTimestamp="2025-11-26 06:45:46 +0000 UTC" firstStartedPulling="2025-11-26 06:45:47.19159577 +0000 UTC m=+1230.552899732" lastFinishedPulling="2025-11-26 06:45:57.969692427 +0000 UTC m=+1241.330996379" observedRunningTime="2025-11-26 06:45:58.928011347 +0000 UTC m=+1242.289315299" watchObservedRunningTime="2025-11-26 06:45:58.935755783 +0000 UTC m=+1242.297059735" Nov 26 06:46:00 crc kubenswrapper[4775]: I1126 06:46:00.651394 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 26 06:46:02 crc kubenswrapper[4775]: I1126 06:46:02.695096 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 26 06:46:11 crc kubenswrapper[4775]: I1126 06:46:11.043062 4775 generic.go:334] "Generic (PLEG): container finished" podID="561d8a6c-3d90-4220-b6ac-386d81f5016f" containerID="88e3d34cd44a362361fab1ffa9bb55208689d59183488afd0a644b02143b8eb7" exitCode=0 Nov 26 06:46:11 crc kubenswrapper[4775]: I1126 06:46:11.043174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" event={"ID":"561d8a6c-3d90-4220-b6ac-386d81f5016f","Type":"ContainerDied","Data":"88e3d34cd44a362361fab1ffa9bb55208689d59183488afd0a644b02143b8eb7"} Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.440827 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.536699 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-repo-setup-combined-ca-bundle\") pod \"561d8a6c-3d90-4220-b6ac-386d81f5016f\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.537061 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-ssh-key\") pod \"561d8a6c-3d90-4220-b6ac-386d81f5016f\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.537361 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-inventory\") pod \"561d8a6c-3d90-4220-b6ac-386d81f5016f\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.537503 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt7nc\" (UniqueName: \"kubernetes.io/projected/561d8a6c-3d90-4220-b6ac-386d81f5016f-kube-api-access-gt7nc\") pod \"561d8a6c-3d90-4220-b6ac-386d81f5016f\" (UID: \"561d8a6c-3d90-4220-b6ac-386d81f5016f\") " Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.543987 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "561d8a6c-3d90-4220-b6ac-386d81f5016f" (UID: "561d8a6c-3d90-4220-b6ac-386d81f5016f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.549035 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561d8a6c-3d90-4220-b6ac-386d81f5016f-kube-api-access-gt7nc" (OuterVolumeSpecName: "kube-api-access-gt7nc") pod "561d8a6c-3d90-4220-b6ac-386d81f5016f" (UID: "561d8a6c-3d90-4220-b6ac-386d81f5016f"). InnerVolumeSpecName "kube-api-access-gt7nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.564158 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-inventory" (OuterVolumeSpecName: "inventory") pod "561d8a6c-3d90-4220-b6ac-386d81f5016f" (UID: "561d8a6c-3d90-4220-b6ac-386d81f5016f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.564521 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "561d8a6c-3d90-4220-b6ac-386d81f5016f" (UID: "561d8a6c-3d90-4220-b6ac-386d81f5016f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.639683 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt7nc\" (UniqueName: \"kubernetes.io/projected/561d8a6c-3d90-4220-b6ac-386d81f5016f-kube-api-access-gt7nc\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.639762 4775 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.639783 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:12 crc kubenswrapper[4775]: I1126 06:46:12.639797 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/561d8a6c-3d90-4220-b6ac-386d81f5016f-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.066143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" event={"ID":"561d8a6c-3d90-4220-b6ac-386d81f5016f","Type":"ContainerDied","Data":"dcff3387ad051119e3ab19438b4c2c6c66c0e906e82c87953af376af3ac9c470"} Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.066457 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcff3387ad051119e3ab19438b4c2c6c66c0e906e82c87953af376af3ac9c470" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.066303 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.172842 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2"] Nov 26 06:46:13 crc kubenswrapper[4775]: E1126 06:46:13.173492 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561d8a6c-3d90-4220-b6ac-386d81f5016f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.173610 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="561d8a6c-3d90-4220-b6ac-386d81f5016f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.173960 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="561d8a6c-3d90-4220-b6ac-386d81f5016f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.174844 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.183069 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2"] Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.217995 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.218160 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.218244 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.218246 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.250097 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.250144 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7gbj\" (UniqueName: \"kubernetes.io/projected/f949db3e-70cb-458b-96d9-9c9f0f1118d8-kube-api-access-z7gbj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.250274 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.352049 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7gbj\" (UniqueName: \"kubernetes.io/projected/f949db3e-70cb-458b-96d9-9c9f0f1118d8-kube-api-access-z7gbj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.352158 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.352366 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.356670 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.359875 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.387894 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7gbj\" (UniqueName: \"kubernetes.io/projected/f949db3e-70cb-458b-96d9-9c9f0f1118d8-kube-api-access-z7gbj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-gm5w2\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:13 crc kubenswrapper[4775]: I1126 06:46:13.548059 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:14 crc kubenswrapper[4775]: I1126 06:46:14.160132 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2"] Nov 26 06:46:15 crc kubenswrapper[4775]: I1126 06:46:15.087676 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" event={"ID":"f949db3e-70cb-458b-96d9-9c9f0f1118d8","Type":"ContainerStarted","Data":"b730f73daf977e82cce1cdbf7c4ef8e8cb36861c13dc3df330772526ff2fc6f5"} Nov 26 06:46:15 crc kubenswrapper[4775]: I1126 06:46:15.088253 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" event={"ID":"f949db3e-70cb-458b-96d9-9c9f0f1118d8","Type":"ContainerStarted","Data":"b4b128ffcee8561585db80f3b35f023fd85b2fa840fa78dde2f38d78a495a49f"} Nov 26 06:46:15 crc kubenswrapper[4775]: I1126 06:46:15.106605 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" podStartSLOduration=1.511919785 podStartE2EDuration="2.106579933s" podCreationTimestamp="2025-11-26 06:46:13 +0000 UTC" firstStartedPulling="2025-11-26 06:46:14.160851478 +0000 UTC m=+1257.522155450" lastFinishedPulling="2025-11-26 06:46:14.755511636 +0000 UTC m=+1258.116815598" observedRunningTime="2025-11-26 06:46:15.10308901 +0000 UTC m=+1258.464392972" watchObservedRunningTime="2025-11-26 06:46:15.106579933 +0000 UTC m=+1258.467883905" Nov 26 06:46:18 crc kubenswrapper[4775]: I1126 06:46:18.124544 4775 generic.go:334] "Generic (PLEG): container finished" podID="f949db3e-70cb-458b-96d9-9c9f0f1118d8" containerID="b730f73daf977e82cce1cdbf7c4ef8e8cb36861c13dc3df330772526ff2fc6f5" exitCode=0 Nov 26 06:46:18 crc kubenswrapper[4775]: I1126 06:46:18.124608 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" event={"ID":"f949db3e-70cb-458b-96d9-9c9f0f1118d8","Type":"ContainerDied","Data":"b730f73daf977e82cce1cdbf7c4ef8e8cb36861c13dc3df330772526ff2fc6f5"} Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.551515 4775 scope.go:117] "RemoveContainer" containerID="d3d8bf50cc4011580997dec0de0a58ddd41324f8d8e9ebe5e2345ec1b7ba292c" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.575134 4775 scope.go:117] "RemoveContainer" containerID="6ca8a1a2707962f20c8b0b03e1c98798df5ec08fcabad6ff58f7fc18121991c7" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.691850 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.789911 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-ssh-key\") pod \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.791235 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-inventory\") pod \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.791368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7gbj\" (UniqueName: \"kubernetes.io/projected/f949db3e-70cb-458b-96d9-9c9f0f1118d8-kube-api-access-z7gbj\") pod \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\" (UID: \"f949db3e-70cb-458b-96d9-9c9f0f1118d8\") " Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.803137 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f949db3e-70cb-458b-96d9-9c9f0f1118d8-kube-api-access-z7gbj" (OuterVolumeSpecName: "kube-api-access-z7gbj") pod "f949db3e-70cb-458b-96d9-9c9f0f1118d8" (UID: "f949db3e-70cb-458b-96d9-9c9f0f1118d8"). InnerVolumeSpecName "kube-api-access-z7gbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.827216 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f949db3e-70cb-458b-96d9-9c9f0f1118d8" (UID: "f949db3e-70cb-458b-96d9-9c9f0f1118d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.828781 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-inventory" (OuterVolumeSpecName: "inventory") pod "f949db3e-70cb-458b-96d9-9c9f0f1118d8" (UID: "f949db3e-70cb-458b-96d9-9c9f0f1118d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.893977 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.894012 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7gbj\" (UniqueName: \"kubernetes.io/projected/f949db3e-70cb-458b-96d9-9c9f0f1118d8-kube-api-access-z7gbj\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:19 crc kubenswrapper[4775]: I1126 06:46:19.894028 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f949db3e-70cb-458b-96d9-9c9f0f1118d8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.151161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" event={"ID":"f949db3e-70cb-458b-96d9-9c9f0f1118d8","Type":"ContainerDied","Data":"b4b128ffcee8561585db80f3b35f023fd85b2fa840fa78dde2f38d78a495a49f"} Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.151211 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4b128ffcee8561585db80f3b35f023fd85b2fa840fa78dde2f38d78a495a49f" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.151224 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-gm5w2" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.230876 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz"] Nov 26 06:46:20 crc kubenswrapper[4775]: E1126 06:46:20.231497 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f949db3e-70cb-458b-96d9-9c9f0f1118d8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.231516 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f949db3e-70cb-458b-96d9-9c9f0f1118d8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.231851 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f949db3e-70cb-458b-96d9-9c9f0f1118d8" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.232497 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.237227 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.237290 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.237415 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.237475 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.242449 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz"] Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.301529 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.301639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzkjq\" (UniqueName: \"kubernetes.io/projected/66abe898-3799-4e9d-abc7-748423d2daa3-kube-api-access-gzkjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.301670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.301697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.403873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.404045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzkjq\" (UniqueName: \"kubernetes.io/projected/66abe898-3799-4e9d-abc7-748423d2daa3-kube-api-access-gzkjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.404082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.404114 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.419401 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.421574 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.425523 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.436558 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzkjq\" (UniqueName: \"kubernetes.io/projected/66abe898-3799-4e9d-abc7-748423d2daa3-kube-api-access-gzkjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:20 crc kubenswrapper[4775]: I1126 06:46:20.553127 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:46:21 crc kubenswrapper[4775]: W1126 06:46:21.111984 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66abe898_3799_4e9d_abc7_748423d2daa3.slice/crio-10d49a3a2287ee0f3a9ac9d2d71cb21fea5cce9e3d172cfab09ed6a3f4d68483 WatchSource:0}: Error finding container 10d49a3a2287ee0f3a9ac9d2d71cb21fea5cce9e3d172cfab09ed6a3f4d68483: Status 404 returned error can't find the container with id 10d49a3a2287ee0f3a9ac9d2d71cb21fea5cce9e3d172cfab09ed6a3f4d68483 Nov 26 06:46:21 crc kubenswrapper[4775]: I1126 06:46:21.118269 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz"] Nov 26 06:46:21 crc kubenswrapper[4775]: I1126 06:46:21.166007 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" event={"ID":"66abe898-3799-4e9d-abc7-748423d2daa3","Type":"ContainerStarted","Data":"10d49a3a2287ee0f3a9ac9d2d71cb21fea5cce9e3d172cfab09ed6a3f4d68483"} Nov 26 06:46:21 crc kubenswrapper[4775]: I1126 06:46:21.420252 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:46:21 crc kubenswrapper[4775]: I1126 06:46:21.420358 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:46:22 crc kubenswrapper[4775]: I1126 06:46:22.176926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" event={"ID":"66abe898-3799-4e9d-abc7-748423d2daa3","Type":"ContainerStarted","Data":"eae3cafe56156d305f31e81fed22e9c98b878f132187c5a6a46b63ebaf162085"} Nov 26 06:46:22 crc kubenswrapper[4775]: I1126 06:46:22.200269 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" podStartSLOduration=1.749651439 podStartE2EDuration="2.200250801s" podCreationTimestamp="2025-11-26 06:46:20 +0000 UTC" firstStartedPulling="2025-11-26 06:46:21.118243875 +0000 UTC m=+1264.479547867" lastFinishedPulling="2025-11-26 06:46:21.568843277 +0000 UTC m=+1264.930147229" observedRunningTime="2025-11-26 06:46:22.195451374 +0000 UTC m=+1265.556755326" watchObservedRunningTime="2025-11-26 06:46:22.200250801 +0000 UTC m=+1265.561554743" Nov 26 06:46:51 crc kubenswrapper[4775]: I1126 06:46:51.420097 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:46:51 crc kubenswrapper[4775]: I1126 06:46:51.420613 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:46:51 crc kubenswrapper[4775]: I1126 06:46:51.420691 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:46:51 crc kubenswrapper[4775]: I1126 06:46:51.421484 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8edde92e1ddfc192e6ae3206422c2d764bb7e89d2f7777b6c406add7b5a50f48"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:46:51 crc kubenswrapper[4775]: I1126 06:46:51.421568 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://8edde92e1ddfc192e6ae3206422c2d764bb7e89d2f7777b6c406add7b5a50f48" gracePeriod=600 Nov 26 06:46:52 crc kubenswrapper[4775]: I1126 06:46:52.515187 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="8edde92e1ddfc192e6ae3206422c2d764bb7e89d2f7777b6c406add7b5a50f48" exitCode=0 Nov 26 06:46:52 crc kubenswrapper[4775]: I1126 06:46:52.515409 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"8edde92e1ddfc192e6ae3206422c2d764bb7e89d2f7777b6c406add7b5a50f48"} Nov 26 06:46:52 crc kubenswrapper[4775]: I1126 06:46:52.515782 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56"} Nov 26 06:46:52 crc kubenswrapper[4775]: I1126 06:46:52.515808 4775 scope.go:117] "RemoveContainer" containerID="8e3a269c3e7b57838eb3de922a5f67cb80c1084327864aee098382792cdcca9f" Nov 26 06:47:19 crc kubenswrapper[4775]: I1126 06:47:19.669921 4775 scope.go:117] "RemoveContainer" containerID="b5ba65b78191c56fd2f5ab03c97b3736920b1b6922f7422278b37c8933da576c" Nov 26 06:48:19 crc kubenswrapper[4775]: I1126 06:48:19.775595 4775 scope.go:117] "RemoveContainer" containerID="f8bcd7ec20eb8a4ae97d80fe889b3e59b2b484ea35b4fbc229a8c5ede4cf3cfb" Nov 26 06:48:51 crc kubenswrapper[4775]: I1126 06:48:51.420040 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:48:51 crc kubenswrapper[4775]: I1126 06:48:51.420651 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:49:21 crc kubenswrapper[4775]: I1126 06:49:21.420088 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:49:21 crc kubenswrapper[4775]: I1126 06:49:21.420887 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:49:26 crc kubenswrapper[4775]: I1126 06:49:26.252811 4775 generic.go:334] "Generic (PLEG): container finished" podID="66abe898-3799-4e9d-abc7-748423d2daa3" containerID="eae3cafe56156d305f31e81fed22e9c98b878f132187c5a6a46b63ebaf162085" exitCode=0 Nov 26 06:49:26 crc kubenswrapper[4775]: I1126 06:49:26.252932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" event={"ID":"66abe898-3799-4e9d-abc7-748423d2daa3","Type":"ContainerDied","Data":"eae3cafe56156d305f31e81fed22e9c98b878f132187c5a6a46b63ebaf162085"} Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.710043 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.732109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-bootstrap-combined-ca-bundle\") pod \"66abe898-3799-4e9d-abc7-748423d2daa3\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.732191 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzkjq\" (UniqueName: \"kubernetes.io/projected/66abe898-3799-4e9d-abc7-748423d2daa3-kube-api-access-gzkjq\") pod \"66abe898-3799-4e9d-abc7-748423d2daa3\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.732248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-inventory\") pod \"66abe898-3799-4e9d-abc7-748423d2daa3\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.732346 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-ssh-key\") pod \"66abe898-3799-4e9d-abc7-748423d2daa3\" (UID: \"66abe898-3799-4e9d-abc7-748423d2daa3\") " Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.741639 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66abe898-3799-4e9d-abc7-748423d2daa3-kube-api-access-gzkjq" (OuterVolumeSpecName: "kube-api-access-gzkjq") pod "66abe898-3799-4e9d-abc7-748423d2daa3" (UID: "66abe898-3799-4e9d-abc7-748423d2daa3"). InnerVolumeSpecName "kube-api-access-gzkjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.752947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "66abe898-3799-4e9d-abc7-748423d2daa3" (UID: "66abe898-3799-4e9d-abc7-748423d2daa3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.784153 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66abe898-3799-4e9d-abc7-748423d2daa3" (UID: "66abe898-3799-4e9d-abc7-748423d2daa3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.792659 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-inventory" (OuterVolumeSpecName: "inventory") pod "66abe898-3799-4e9d-abc7-748423d2daa3" (UID: "66abe898-3799-4e9d-abc7-748423d2daa3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.835178 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.835253 4775 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.835270 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzkjq\" (UniqueName: \"kubernetes.io/projected/66abe898-3799-4e9d-abc7-748423d2daa3-kube-api-access-gzkjq\") on node \"crc\" DevicePath \"\"" Nov 26 06:49:27 crc kubenswrapper[4775]: I1126 06:49:27.835283 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66abe898-3799-4e9d-abc7-748423d2daa3-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.280686 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" event={"ID":"66abe898-3799-4e9d-abc7-748423d2daa3","Type":"ContainerDied","Data":"10d49a3a2287ee0f3a9ac9d2d71cb21fea5cce9e3d172cfab09ed6a3f4d68483"} Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.280785 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10d49a3a2287ee0f3a9ac9d2d71cb21fea5cce9e3d172cfab09ed6a3f4d68483" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.280844 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.398524 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh"] Nov 26 06:49:28 crc kubenswrapper[4775]: E1126 06:49:28.399013 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66abe898-3799-4e9d-abc7-748423d2daa3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.399036 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="66abe898-3799-4e9d-abc7-748423d2daa3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.399297 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="66abe898-3799-4e9d-abc7-748423d2daa3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.400363 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.403886 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.404028 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.404423 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.409872 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.421996 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh"] Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.456601 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.456748 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnhwh\" (UniqueName: \"kubernetes.io/projected/e780accc-4808-46d6-b5df-5b6492107f25-kube-api-access-mnhwh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.456872 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.557956 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnhwh\" (UniqueName: \"kubernetes.io/projected/e780accc-4808-46d6-b5df-5b6492107f25-kube-api-access-mnhwh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.558116 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.558197 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.563053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.564997 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.575146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnhwh\" (UniqueName: \"kubernetes.io/projected/e780accc-4808-46d6-b5df-5b6492107f25-kube-api-access-mnhwh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:28 crc kubenswrapper[4775]: I1126 06:49:28.738904 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:49:29 crc kubenswrapper[4775]: I1126 06:49:29.353065 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh"] Nov 26 06:49:30 crc kubenswrapper[4775]: I1126 06:49:30.304004 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" event={"ID":"e780accc-4808-46d6-b5df-5b6492107f25","Type":"ContainerStarted","Data":"852435d47dc49acc9a9e65bb7d9107d091f41193c974a27170b7311823712a8c"} Nov 26 06:49:30 crc kubenswrapper[4775]: I1126 06:49:30.304289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" event={"ID":"e780accc-4808-46d6-b5df-5b6492107f25","Type":"ContainerStarted","Data":"9b4365acc063a15dd8664ec941a49a803d60bc55014ec3702a87251cebfc064d"} Nov 26 06:49:30 crc kubenswrapper[4775]: I1126 06:49:30.335656 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" podStartSLOduration=1.850433387 podStartE2EDuration="2.335635158s" podCreationTimestamp="2025-11-26 06:49:28 +0000 UTC" firstStartedPulling="2025-11-26 06:49:29.354699125 +0000 UTC m=+1452.716003117" lastFinishedPulling="2025-11-26 06:49:29.839900896 +0000 UTC m=+1453.201204888" observedRunningTime="2025-11-26 06:49:30.327899561 +0000 UTC m=+1453.689203503" watchObservedRunningTime="2025-11-26 06:49:30.335635158 +0000 UTC m=+1453.696939120" Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.419928 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.420666 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.420759 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.421743 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.421816 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" gracePeriod=600 Nov 26 06:49:51 crc kubenswrapper[4775]: E1126 06:49:51.548439 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.586374 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" exitCode=0 Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.586428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56"} Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.586468 4775 scope.go:117] "RemoveContainer" containerID="8edde92e1ddfc192e6ae3206422c2d764bb7e89d2f7777b6c406add7b5a50f48" Nov 26 06:49:51 crc kubenswrapper[4775]: I1126 06:49:51.587192 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:49:51 crc kubenswrapper[4775]: E1126 06:49:51.587459 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:50:06 crc kubenswrapper[4775]: I1126 06:50:06.328586 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:50:06 crc kubenswrapper[4775]: E1126 06:50:06.329476 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:50:19 crc kubenswrapper[4775]: I1126 06:50:19.899279 4775 scope.go:117] "RemoveContainer" containerID="20a19d0b0eb11331e934143d7cff3572c583c8581e65550c8b8a42cce159863a" Nov 26 06:50:19 crc kubenswrapper[4775]: I1126 06:50:19.928761 4775 scope.go:117] "RemoveContainer" containerID="8b2efc0704e9f5003ff0b19894d0861682d44396a6b6b8c21a3b29a2c7fef314" Nov 26 06:50:20 crc kubenswrapper[4775]: I1126 06:50:20.328503 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:50:20 crc kubenswrapper[4775]: E1126 06:50:20.329371 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:50:35 crc kubenswrapper[4775]: I1126 06:50:35.328402 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:50:35 crc kubenswrapper[4775]: E1126 06:50:35.329247 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:50:36 crc kubenswrapper[4775]: I1126 06:50:36.070733 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b814-account-create-update-pn25p"] Nov 26 06:50:36 crc kubenswrapper[4775]: I1126 06:50:36.082359 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-w2kd5"] Nov 26 06:50:36 crc kubenswrapper[4775]: I1126 06:50:36.091353 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b814-account-create-update-pn25p"] Nov 26 06:50:36 crc kubenswrapper[4775]: I1126 06:50:36.101317 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-w2kd5"] Nov 26 06:50:37 crc kubenswrapper[4775]: I1126 06:50:37.362647 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e128aee-1376-4ee6-9bea-f5ff40e422d0" path="/var/lib/kubelet/pods/6e128aee-1376-4ee6-9bea-f5ff40e422d0/volumes" Nov 26 06:50:37 crc kubenswrapper[4775]: I1126 06:50:37.365242 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aade9b70-d511-45e4-afcb-9887451cb4ed" path="/var/lib/kubelet/pods/aade9b70-d511-45e4-afcb-9887451cb4ed/volumes" Nov 26 06:50:41 crc kubenswrapper[4775]: I1126 06:50:41.028975 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5sdcc"] Nov 26 06:50:41 crc kubenswrapper[4775]: I1126 06:50:41.039183 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5sdcc"] Nov 26 06:50:41 crc kubenswrapper[4775]: I1126 06:50:41.349971 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c54287b0-da13-43bc-9a94-d5fa303b885e" path="/var/lib/kubelet/pods/c54287b0-da13-43bc-9a94-d5fa303b885e/volumes" Nov 26 06:50:42 crc kubenswrapper[4775]: I1126 06:50:42.050818 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-097e-account-create-update-khn5j"] Nov 26 06:50:42 crc kubenswrapper[4775]: I1126 06:50:42.066004 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g6rtl"] Nov 26 06:50:42 crc kubenswrapper[4775]: I1126 06:50:42.078779 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-097e-account-create-update-khn5j"] Nov 26 06:50:42 crc kubenswrapper[4775]: I1126 06:50:42.090887 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g6rtl"] Nov 26 06:50:42 crc kubenswrapper[4775]: I1126 06:50:42.103642 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b988-account-create-update-7tg4l"] Nov 26 06:50:42 crc kubenswrapper[4775]: I1126 06:50:42.112041 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b988-account-create-update-7tg4l"] Nov 26 06:50:43 crc kubenswrapper[4775]: I1126 06:50:43.339818 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73081230-d192-4993-9434-9c4c6a5533f9" path="/var/lib/kubelet/pods/73081230-d192-4993-9434-9c4c6a5533f9/volumes" Nov 26 06:50:43 crc kubenswrapper[4775]: I1126 06:50:43.340963 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8957b6f-3c9b-4562-9720-c9dbc09d3ee2" path="/var/lib/kubelet/pods/b8957b6f-3c9b-4562-9720-c9dbc09d3ee2/volumes" Nov 26 06:50:43 crc kubenswrapper[4775]: I1126 06:50:43.341579 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab" path="/var/lib/kubelet/pods/e4fa84d7-3f4b-4ac4-a0b5-c9498f33e3ab/volumes" Nov 26 06:50:46 crc kubenswrapper[4775]: I1126 06:50:46.327908 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:50:46 crc kubenswrapper[4775]: E1126 06:50:46.329300 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:50:59 crc kubenswrapper[4775]: I1126 06:50:59.328671 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:50:59 crc kubenswrapper[4775]: E1126 06:50:59.329965 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.056452 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-bnfj2"] Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.068128 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-bnfj2"] Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.396324 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z7jtd"] Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.400847 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.415395 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7jtd"] Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.452901 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-utilities\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.452963 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wh5v\" (UniqueName: \"kubernetes.io/projected/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-kube-api-access-8wh5v\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.453060 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-catalog-content\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.554383 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-catalog-content\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.554506 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-utilities\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.554530 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wh5v\" (UniqueName: \"kubernetes.io/projected/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-kube-api-access-8wh5v\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.555343 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-utilities\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.555454 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-catalog-content\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.575344 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wh5v\" (UniqueName: \"kubernetes.io/projected/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-kube-api-access-8wh5v\") pod \"redhat-operators-z7jtd\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:00 crc kubenswrapper[4775]: I1126 06:51:00.736293 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:01 crc kubenswrapper[4775]: I1126 06:51:01.265094 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7jtd"] Nov 26 06:51:01 crc kubenswrapper[4775]: I1126 06:51:01.338610 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06776a0e-14fb-411c-b70d-861de9969bcd" path="/var/lib/kubelet/pods/06776a0e-14fb-411c-b70d-861de9969bcd/volumes" Nov 26 06:51:01 crc kubenswrapper[4775]: I1126 06:51:01.375666 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7jtd" event={"ID":"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2","Type":"ContainerStarted","Data":"a9bda9fbba5524d736db7a45878a0e3154f29408c6c49593a1eb682dfe27e411"} Nov 26 06:51:02 crc kubenswrapper[4775]: I1126 06:51:02.391253 4775 generic.go:334] "Generic (PLEG): container finished" podID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerID="1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5" exitCode=0 Nov 26 06:51:02 crc kubenswrapper[4775]: I1126 06:51:02.391367 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7jtd" event={"ID":"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2","Type":"ContainerDied","Data":"1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5"} Nov 26 06:51:02 crc kubenswrapper[4775]: I1126 06:51:02.396605 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 06:51:04 crc kubenswrapper[4775]: I1126 06:51:04.422532 4775 generic.go:334] "Generic (PLEG): container finished" podID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerID="10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f" exitCode=0 Nov 26 06:51:04 crc kubenswrapper[4775]: I1126 06:51:04.422640 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7jtd" event={"ID":"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2","Type":"ContainerDied","Data":"10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f"} Nov 26 06:51:04 crc kubenswrapper[4775]: I1126 06:51:04.993330 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2lmb2"] Nov 26 06:51:04 crc kubenswrapper[4775]: I1126 06:51:04.996157 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.015388 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lmb2"] Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.044507 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-utilities\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.044560 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkxrt\" (UniqueName: \"kubernetes.io/projected/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-kube-api-access-xkxrt\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.044846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-catalog-content\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.146903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-utilities\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.146974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkxrt\" (UniqueName: \"kubernetes.io/projected/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-kube-api-access-xkxrt\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.147050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-catalog-content\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.147632 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-catalog-content\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.147746 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-utilities\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.171882 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkxrt\" (UniqueName: \"kubernetes.io/projected/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-kube-api-access-xkxrt\") pod \"redhat-marketplace-2lmb2\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.361491 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.456263 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7jtd" event={"ID":"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2","Type":"ContainerStarted","Data":"68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105"} Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.458645 4775 generic.go:334] "Generic (PLEG): container finished" podID="e780accc-4808-46d6-b5df-5b6492107f25" containerID="852435d47dc49acc9a9e65bb7d9107d091f41193c974a27170b7311823712a8c" exitCode=0 Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.458684 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" event={"ID":"e780accc-4808-46d6-b5df-5b6492107f25","Type":"ContainerDied","Data":"852435d47dc49acc9a9e65bb7d9107d091f41193c974a27170b7311823712a8c"} Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.516301 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z7jtd" podStartSLOduration=3.043529183 podStartE2EDuration="5.516277177s" podCreationTimestamp="2025-11-26 06:51:00 +0000 UTC" firstStartedPulling="2025-11-26 06:51:02.396185765 +0000 UTC m=+1545.757489757" lastFinishedPulling="2025-11-26 06:51:04.868933799 +0000 UTC m=+1548.230237751" observedRunningTime="2025-11-26 06:51:05.489320347 +0000 UTC m=+1548.850624299" watchObservedRunningTime="2025-11-26 06:51:05.516277177 +0000 UTC m=+1548.877581139" Nov 26 06:51:05 crc kubenswrapper[4775]: I1126 06:51:05.861552 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lmb2"] Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.478379 4775 generic.go:334] "Generic (PLEG): container finished" podID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerID="49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300" exitCode=0 Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.478574 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lmb2" event={"ID":"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0","Type":"ContainerDied","Data":"49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300"} Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.478750 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lmb2" event={"ID":"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0","Type":"ContainerStarted","Data":"dbbc3aa4d5fed8993538bb5765d74fadc9e2e30819538e0cc50d7013cdb0bef5"} Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.921115 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.988378 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnhwh\" (UniqueName: \"kubernetes.io/projected/e780accc-4808-46d6-b5df-5b6492107f25-kube-api-access-mnhwh\") pod \"e780accc-4808-46d6-b5df-5b6492107f25\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.988465 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-inventory\") pod \"e780accc-4808-46d6-b5df-5b6492107f25\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " Nov 26 06:51:06 crc kubenswrapper[4775]: I1126 06:51:06.988751 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-ssh-key\") pod \"e780accc-4808-46d6-b5df-5b6492107f25\" (UID: \"e780accc-4808-46d6-b5df-5b6492107f25\") " Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.003077 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e780accc-4808-46d6-b5df-5b6492107f25-kube-api-access-mnhwh" (OuterVolumeSpecName: "kube-api-access-mnhwh") pod "e780accc-4808-46d6-b5df-5b6492107f25" (UID: "e780accc-4808-46d6-b5df-5b6492107f25"). InnerVolumeSpecName "kube-api-access-mnhwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.030199 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e780accc-4808-46d6-b5df-5b6492107f25" (UID: "e780accc-4808-46d6-b5df-5b6492107f25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.030075 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-inventory" (OuterVolumeSpecName: "inventory") pod "e780accc-4808-46d6-b5df-5b6492107f25" (UID: "e780accc-4808-46d6-b5df-5b6492107f25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.091324 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnhwh\" (UniqueName: \"kubernetes.io/projected/e780accc-4808-46d6-b5df-5b6492107f25-kube-api-access-mnhwh\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.091354 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.091363 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e780accc-4808-46d6-b5df-5b6492107f25-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.489261 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" event={"ID":"e780accc-4808-46d6-b5df-5b6492107f25","Type":"ContainerDied","Data":"9b4365acc063a15dd8664ec941a49a803d60bc55014ec3702a87251cebfc064d"} Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.489641 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b4365acc063a15dd8664ec941a49a803d60bc55014ec3702a87251cebfc064d" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.489494 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.492318 4775 generic.go:334] "Generic (PLEG): container finished" podID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerID="ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5" exitCode=0 Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.492373 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lmb2" event={"ID":"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0","Type":"ContainerDied","Data":"ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5"} Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.590740 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k"] Nov 26 06:51:07 crc kubenswrapper[4775]: E1126 06:51:07.591446 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e780accc-4808-46d6-b5df-5b6492107f25" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.591470 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e780accc-4808-46d6-b5df-5b6492107f25" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.591788 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e780accc-4808-46d6-b5df-5b6492107f25" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.592452 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.594559 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.594776 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.594956 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.595262 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.614833 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k"] Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.703277 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.703418 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.703458 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcs6q\" (UniqueName: \"kubernetes.io/projected/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-kube-api-access-bcs6q\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.806770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.806875 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.806903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcs6q\" (UniqueName: \"kubernetes.io/projected/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-kube-api-access-bcs6q\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.819559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.831956 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.848415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcs6q\" (UniqueName: \"kubernetes.io/projected/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-kube-api-access-bcs6q\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kp72k\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:07 crc kubenswrapper[4775]: I1126 06:51:07.918393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:51:08 crc kubenswrapper[4775]: I1126 06:51:08.488216 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k"] Nov 26 06:51:08 crc kubenswrapper[4775]: I1126 06:51:08.518004 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" event={"ID":"7b826be9-41b3-4715-b7cb-5b8bc10e35a9","Type":"ContainerStarted","Data":"b010a1a3d26be69697e5782780318fc68e7de7d11de2852392aa3ecfbc5b09f8"} Nov 26 06:51:08 crc kubenswrapper[4775]: I1126 06:51:08.523158 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lmb2" event={"ID":"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0","Type":"ContainerStarted","Data":"21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231"} Nov 26 06:51:08 crc kubenswrapper[4775]: I1126 06:51:08.550173 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2lmb2" podStartSLOduration=3.106269398 podStartE2EDuration="4.550153268s" podCreationTimestamp="2025-11-26 06:51:04 +0000 UTC" firstStartedPulling="2025-11-26 06:51:06.482054294 +0000 UTC m=+1549.843358246" lastFinishedPulling="2025-11-26 06:51:07.925938164 +0000 UTC m=+1551.287242116" observedRunningTime="2025-11-26 06:51:08.543730389 +0000 UTC m=+1551.905034351" watchObservedRunningTime="2025-11-26 06:51:08.550153268 +0000 UTC m=+1551.911457220" Nov 26 06:51:09 crc kubenswrapper[4775]: I1126 06:51:09.535855 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" event={"ID":"7b826be9-41b3-4715-b7cb-5b8bc10e35a9","Type":"ContainerStarted","Data":"9210c36c4f40dcc257089d611c183e36e0c0ce11de30877b5939098a710c932c"} Nov 26 06:51:09 crc kubenswrapper[4775]: I1126 06:51:09.563732 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" podStartSLOduration=1.893864041 podStartE2EDuration="2.563693765s" podCreationTimestamp="2025-11-26 06:51:07 +0000 UTC" firstStartedPulling="2025-11-26 06:51:08.493046818 +0000 UTC m=+1551.854350770" lastFinishedPulling="2025-11-26 06:51:09.162876532 +0000 UTC m=+1552.524180494" observedRunningTime="2025-11-26 06:51:09.552167164 +0000 UTC m=+1552.913471136" watchObservedRunningTime="2025-11-26 06:51:09.563693765 +0000 UTC m=+1552.924997727" Nov 26 06:51:10 crc kubenswrapper[4775]: I1126 06:51:10.739623 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:10 crc kubenswrapper[4775]: I1126 06:51:10.739697 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:11 crc kubenswrapper[4775]: I1126 06:51:11.786201 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z7jtd" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="registry-server" probeResult="failure" output=< Nov 26 06:51:11 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Nov 26 06:51:11 crc kubenswrapper[4775]: > Nov 26 06:51:12 crc kubenswrapper[4775]: I1126 06:51:12.328333 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:51:12 crc kubenswrapper[4775]: E1126 06:51:12.328993 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.036551 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c40d-account-create-update-snvdn"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.046771 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-dd94-account-create-update-7zjm6"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.056108 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-fkk8f"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.068409 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-w8tnl"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.075994 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c40d-account-create-update-snvdn"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.083321 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-j4rs9"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.090228 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c371-account-create-update-jt4mk"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.099820 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-fkk8f"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.108255 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-dd94-account-create-update-7zjm6"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.115821 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-j4rs9"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.122938 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-w8tnl"] Nov 26 06:51:14 crc kubenswrapper[4775]: I1126 06:51:14.129639 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c371-account-create-update-jt4mk"] Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.342528 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="229de831-b61a-4cee-8005-5a53eede49a1" path="/var/lib/kubelet/pods/229de831-b61a-4cee-8005-5a53eede49a1/volumes" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.343274 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="759a51cc-5fd3-4c7d-bfd8-072222f1e2e5" path="/var/lib/kubelet/pods/759a51cc-5fd3-4c7d-bfd8-072222f1e2e5/volumes" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.343834 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d56ad5c-f3dc-4d36-862b-a68529cffacb" path="/var/lib/kubelet/pods/8d56ad5c-f3dc-4d36-862b-a68529cffacb/volumes" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.344371 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c651662-819c-4691-9632-5948a2b3d88c" path="/var/lib/kubelet/pods/9c651662-819c-4691-9632-5948a2b3d88c/volumes" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.345511 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e342a0-cce3-4690-93ef-f1afd6f8e44a" path="/var/lib/kubelet/pods/a7e342a0-cce3-4690-93ef-f1afd6f8e44a/volumes" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.346343 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac40cdd2-dadd-43ee-bf2f-d6df962d5d52" path="/var/lib/kubelet/pods/ac40cdd2-dadd-43ee-bf2f-d6df962d5d52/volumes" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.362622 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.364046 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.415338 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.668775 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:15 crc kubenswrapper[4775]: I1126 06:51:15.726130 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lmb2"] Nov 26 06:51:17 crc kubenswrapper[4775]: I1126 06:51:17.625823 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2lmb2" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="registry-server" containerID="cri-o://21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231" gracePeriod=2 Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.146739 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.322628 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-utilities\") pod \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.322703 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-catalog-content\") pod \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.322855 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkxrt\" (UniqueName: \"kubernetes.io/projected/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-kube-api-access-xkxrt\") pod \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\" (UID: \"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0\") " Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.325904 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-utilities" (OuterVolumeSpecName: "utilities") pod "c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" (UID: "c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.330518 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-kube-api-access-xkxrt" (OuterVolumeSpecName: "kube-api-access-xkxrt") pod "c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" (UID: "c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0"). InnerVolumeSpecName "kube-api-access-xkxrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.350502 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" (UID: "c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.425537 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.425567 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.425577 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkxrt\" (UniqueName: \"kubernetes.io/projected/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0-kube-api-access-xkxrt\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.637578 4775 generic.go:334] "Generic (PLEG): container finished" podID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerID="21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231" exitCode=0 Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.637623 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lmb2" event={"ID":"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0","Type":"ContainerDied","Data":"21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231"} Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.637665 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2lmb2" event={"ID":"c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0","Type":"ContainerDied","Data":"dbbc3aa4d5fed8993538bb5765d74fadc9e2e30819538e0cc50d7013cdb0bef5"} Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.637675 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2lmb2" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.637687 4775 scope.go:117] "RemoveContainer" containerID="21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.670585 4775 scope.go:117] "RemoveContainer" containerID="ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.674281 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lmb2"] Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.681677 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2lmb2"] Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.720061 4775 scope.go:117] "RemoveContainer" containerID="49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.745347 4775 scope.go:117] "RemoveContainer" containerID="21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231" Nov 26 06:51:18 crc kubenswrapper[4775]: E1126 06:51:18.745814 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231\": container with ID starting with 21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231 not found: ID does not exist" containerID="21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.745857 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231"} err="failed to get container status \"21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231\": rpc error: code = NotFound desc = could not find container \"21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231\": container with ID starting with 21f65b86f061862b43fdf19b34d54d0a363c51fe3289dde9fff99657d48e6231 not found: ID does not exist" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.745885 4775 scope.go:117] "RemoveContainer" containerID="ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5" Nov 26 06:51:18 crc kubenswrapper[4775]: E1126 06:51:18.746162 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5\": container with ID starting with ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5 not found: ID does not exist" containerID="ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.746212 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5"} err="failed to get container status \"ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5\": rpc error: code = NotFound desc = could not find container \"ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5\": container with ID starting with ead39fa7744870574f4325212410c8404e77461036541aaddc816398b7ed44d5 not found: ID does not exist" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.746247 4775 scope.go:117] "RemoveContainer" containerID="49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300" Nov 26 06:51:18 crc kubenswrapper[4775]: E1126 06:51:18.746546 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300\": container with ID starting with 49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300 not found: ID does not exist" containerID="49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300" Nov 26 06:51:18 crc kubenswrapper[4775]: I1126 06:51:18.746572 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300"} err="failed to get container status \"49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300\": rpc error: code = NotFound desc = could not find container \"49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300\": container with ID starting with 49dcf5f605929e7ecea25dbbac806044e0c4fcba82d59bd0225103f0b6b4c300 not found: ID does not exist" Nov 26 06:51:19 crc kubenswrapper[4775]: I1126 06:51:19.063480 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-7kcv8"] Nov 26 06:51:19 crc kubenswrapper[4775]: I1126 06:51:19.072454 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-7kcv8"] Nov 26 06:51:19 crc kubenswrapper[4775]: I1126 06:51:19.340197 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea8f43d-f87b-4f26-9d09-465c46a9e853" path="/var/lib/kubelet/pods/1ea8f43d-f87b-4f26-9d09-465c46a9e853/volumes" Nov 26 06:51:19 crc kubenswrapper[4775]: I1126 06:51:19.341122 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" path="/var/lib/kubelet/pods/c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0/volumes" Nov 26 06:51:19 crc kubenswrapper[4775]: I1126 06:51:19.999934 4775 scope.go:117] "RemoveContainer" containerID="4bba4b9188b86331293b68e54a72fe7c36d900aa577e884627b2e9888a7e19ec" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.028102 4775 scope.go:117] "RemoveContainer" containerID="4383c34721635f3ef8676856d6f0165c4adeb93596757a9dd69c46556d6ad68c" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.097891 4775 scope.go:117] "RemoveContainer" containerID="ce0ab6e4d4886a95817154fa408d14227f06457a00a04fc5506c23eca7758e63" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.149544 4775 scope.go:117] "RemoveContainer" containerID="b698c78b7fb72307ce47dcf4e5ae0aeb354e3604c1148cc1a8353503afb6394e" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.180918 4775 scope.go:117] "RemoveContainer" containerID="7c4e4a3df95422486a05e3aa000604a4931381c03330e4bfbaa59a7484d49f93" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.229541 4775 scope.go:117] "RemoveContainer" containerID="163fefffded403580198d0723221e6aef8818363230465cdd4a6e640b2e9ef3a" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.272890 4775 scope.go:117] "RemoveContainer" containerID="53b8e58d4dccb836949653f67bce2bf8bad8e30359f230e26554bf8609f65956" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.292159 4775 scope.go:117] "RemoveContainer" containerID="6c51504d44cfa68b422120558f2b17abc586f5fabec98b08f856dc57e5118dab" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.314819 4775 scope.go:117] "RemoveContainer" containerID="6b41a0b963dfac26b8b087dfba156282ed18e588d8998ac10ded3a2f100c4d6d" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.337820 4775 scope.go:117] "RemoveContainer" containerID="058f3a49e4a2725fc75468f5bd01076a3782b831bc88a996b92b3f5c6350b24d" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.366287 4775 scope.go:117] "RemoveContainer" containerID="6e0c23254a48848983f3df52b3e0dc6527463727d8d184c9322ed8149fef90ff" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.391112 4775 scope.go:117] "RemoveContainer" containerID="0e37d7d74b50450e83ffb9050af536a4e36a5438d7290eb9958067ab87c28913" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.416902 4775 scope.go:117] "RemoveContainer" containerID="92412a8829994c6bfbd9ee61f3fc8c92186e5345eb38bfadade4767032bc9d14" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.450522 4775 scope.go:117] "RemoveContainer" containerID="53d27861306ed497ddfe36a93ebdd902141ad3b9522aa0335274ad3b79657c7c" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.789650 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:20 crc kubenswrapper[4775]: I1126 06:51:20.861333 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:22 crc kubenswrapper[4775]: I1126 06:51:22.072426 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z7jtd"] Nov 26 06:51:22 crc kubenswrapper[4775]: I1126 06:51:22.711053 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z7jtd" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="registry-server" containerID="cri-o://68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105" gracePeriod=2 Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.189048 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.318734 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-utilities\") pod \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.318912 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-catalog-content\") pod \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.318947 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wh5v\" (UniqueName: \"kubernetes.io/projected/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-kube-api-access-8wh5v\") pod \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\" (UID: \"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2\") " Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.319965 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-utilities" (OuterVolumeSpecName: "utilities") pod "71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" (UID: "71094c82-3dfe-4dd7-80e6-ac8e340c3ab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.325553 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-kube-api-access-8wh5v" (OuterVolumeSpecName: "kube-api-access-8wh5v") pod "71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" (UID: "71094c82-3dfe-4dd7-80e6-ac8e340c3ab2"). InnerVolumeSpecName "kube-api-access-8wh5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.421658 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.421688 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wh5v\" (UniqueName: \"kubernetes.io/projected/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-kube-api-access-8wh5v\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.487846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" (UID: "71094c82-3dfe-4dd7-80e6-ac8e340c3ab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.523815 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.726852 4775 generic.go:334] "Generic (PLEG): container finished" podID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerID="68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105" exitCode=0 Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.726904 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7jtd" event={"ID":"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2","Type":"ContainerDied","Data":"68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105"} Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.726940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7jtd" event={"ID":"71094c82-3dfe-4dd7-80e6-ac8e340c3ab2","Type":"ContainerDied","Data":"a9bda9fbba5524d736db7a45878a0e3154f29408c6c49593a1eb682dfe27e411"} Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.726944 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7jtd" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.726961 4775 scope.go:117] "RemoveContainer" containerID="68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.763547 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z7jtd"] Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.765059 4775 scope.go:117] "RemoveContainer" containerID="10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.777265 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z7jtd"] Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.787917 4775 scope.go:117] "RemoveContainer" containerID="1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.850635 4775 scope.go:117] "RemoveContainer" containerID="68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105" Nov 26 06:51:23 crc kubenswrapper[4775]: E1126 06:51:23.851205 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105\": container with ID starting with 68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105 not found: ID does not exist" containerID="68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.851240 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105"} err="failed to get container status \"68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105\": rpc error: code = NotFound desc = could not find container \"68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105\": container with ID starting with 68a0342853d69ed31997de35bbae3adc02a844dc8383eb2c66d48f59063a8105 not found: ID does not exist" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.851265 4775 scope.go:117] "RemoveContainer" containerID="10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f" Nov 26 06:51:23 crc kubenswrapper[4775]: E1126 06:51:23.851658 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f\": container with ID starting with 10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f not found: ID does not exist" containerID="10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.851684 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f"} err="failed to get container status \"10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f\": rpc error: code = NotFound desc = could not find container \"10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f\": container with ID starting with 10b09a655e3b8d253abd54b49adf53b88d7bb501413132a87ec48b8decac365f not found: ID does not exist" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.851702 4775 scope.go:117] "RemoveContainer" containerID="1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5" Nov 26 06:51:23 crc kubenswrapper[4775]: E1126 06:51:23.852201 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5\": container with ID starting with 1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5 not found: ID does not exist" containerID="1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5" Nov 26 06:51:23 crc kubenswrapper[4775]: I1126 06:51:23.852249 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5"} err="failed to get container status \"1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5\": rpc error: code = NotFound desc = could not find container \"1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5\": container with ID starting with 1c7ab7a1cd843845d98bd23c0f1f42f2acad97e37bf523e7ffe74803d3a8bbc5 not found: ID does not exist" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.345489 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" path="/var/lib/kubelet/pods/71094c82-3dfe-4dd7-80e6-ac8e340c3ab2/volumes" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.485405 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8x9r6"] Nov 26 06:51:25 crc kubenswrapper[4775]: E1126 06:51:25.486121 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="registry-server" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486144 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="registry-server" Nov 26 06:51:25 crc kubenswrapper[4775]: E1126 06:51:25.486177 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="registry-server" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486190 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="registry-server" Nov 26 06:51:25 crc kubenswrapper[4775]: E1126 06:51:25.486221 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="extract-content" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486235 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="extract-content" Nov 26 06:51:25 crc kubenswrapper[4775]: E1126 06:51:25.486257 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="extract-utilities" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486268 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="extract-utilities" Nov 26 06:51:25 crc kubenswrapper[4775]: E1126 06:51:25.486279 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="extract-content" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486287 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="extract-content" Nov 26 06:51:25 crc kubenswrapper[4775]: E1126 06:51:25.486305 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="extract-utilities" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486313 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="extract-utilities" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486551 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="71094c82-3dfe-4dd7-80e6-ac8e340c3ab2" containerName="registry-server" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.486578 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c92b33d8-6bc0-42d4-9aaa-a0b291ccb9e0" containerName="registry-server" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.489593 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.501842 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8x9r6"] Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.561641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-utilities\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.561798 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvrdk\" (UniqueName: \"kubernetes.io/projected/dcb3b6e8-ca88-4adf-850b-d167527d15d3-kube-api-access-lvrdk\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.561829 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-catalog-content\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.663812 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-utilities\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.663967 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvrdk\" (UniqueName: \"kubernetes.io/projected/dcb3b6e8-ca88-4adf-850b-d167527d15d3-kube-api-access-lvrdk\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.663997 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-catalog-content\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.664366 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-utilities\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.664453 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-catalog-content\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.684272 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvrdk\" (UniqueName: \"kubernetes.io/projected/dcb3b6e8-ca88-4adf-850b-d167527d15d3-kube-api-access-lvrdk\") pod \"community-operators-8x9r6\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:25 crc kubenswrapper[4775]: I1126 06:51:25.831110 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:26 crc kubenswrapper[4775]: I1126 06:51:26.327515 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:51:26 crc kubenswrapper[4775]: E1126 06:51:26.328120 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:51:26 crc kubenswrapper[4775]: I1126 06:51:26.444120 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8x9r6"] Nov 26 06:51:26 crc kubenswrapper[4775]: W1126 06:51:26.445391 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcb3b6e8_ca88_4adf_850b_d167527d15d3.slice/crio-00f3f36a40004d08a274363f69630267f36788bd8fdfc6fd5a98d5b4f9aa959e WatchSource:0}: Error finding container 00f3f36a40004d08a274363f69630267f36788bd8fdfc6fd5a98d5b4f9aa959e: Status 404 returned error can't find the container with id 00f3f36a40004d08a274363f69630267f36788bd8fdfc6fd5a98d5b4f9aa959e Nov 26 06:51:26 crc kubenswrapper[4775]: I1126 06:51:26.762622 4775 generic.go:334] "Generic (PLEG): container finished" podID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerID="c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e" exitCode=0 Nov 26 06:51:26 crc kubenswrapper[4775]: I1126 06:51:26.762689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x9r6" event={"ID":"dcb3b6e8-ca88-4adf-850b-d167527d15d3","Type":"ContainerDied","Data":"c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e"} Nov 26 06:51:26 crc kubenswrapper[4775]: I1126 06:51:26.762783 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x9r6" event={"ID":"dcb3b6e8-ca88-4adf-850b-d167527d15d3","Type":"ContainerStarted","Data":"00f3f36a40004d08a274363f69630267f36788bd8fdfc6fd5a98d5b4f9aa959e"} Nov 26 06:51:28 crc kubenswrapper[4775]: I1126 06:51:28.787236 4775 generic.go:334] "Generic (PLEG): container finished" podID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerID="13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b" exitCode=0 Nov 26 06:51:28 crc kubenswrapper[4775]: I1126 06:51:28.787641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x9r6" event={"ID":"dcb3b6e8-ca88-4adf-850b-d167527d15d3","Type":"ContainerDied","Data":"13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b"} Nov 26 06:51:29 crc kubenswrapper[4775]: I1126 06:51:29.799243 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x9r6" event={"ID":"dcb3b6e8-ca88-4adf-850b-d167527d15d3","Type":"ContainerStarted","Data":"7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8"} Nov 26 06:51:29 crc kubenswrapper[4775]: I1126 06:51:29.826756 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8x9r6" podStartSLOduration=2.299456336 podStartE2EDuration="4.826731124s" podCreationTimestamp="2025-11-26 06:51:25 +0000 UTC" firstStartedPulling="2025-11-26 06:51:26.764561451 +0000 UTC m=+1570.125865423" lastFinishedPulling="2025-11-26 06:51:29.291836239 +0000 UTC m=+1572.653140211" observedRunningTime="2025-11-26 06:51:29.820073581 +0000 UTC m=+1573.181377553" watchObservedRunningTime="2025-11-26 06:51:29.826731124 +0000 UTC m=+1573.188035086" Nov 26 06:51:35 crc kubenswrapper[4775]: I1126 06:51:35.831620 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:35 crc kubenswrapper[4775]: I1126 06:51:35.832216 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:35 crc kubenswrapper[4775]: I1126 06:51:35.897653 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:35 crc kubenswrapper[4775]: I1126 06:51:35.960799 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:36 crc kubenswrapper[4775]: I1126 06:51:36.194303 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8x9r6"] Nov 26 06:51:37 crc kubenswrapper[4775]: I1126 06:51:37.894373 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8x9r6" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="registry-server" containerID="cri-o://7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8" gracePeriod=2 Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.327887 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:51:38 crc kubenswrapper[4775]: E1126 06:51:38.328552 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.371297 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.432005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-catalog-content\") pod \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.432081 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvrdk\" (UniqueName: \"kubernetes.io/projected/dcb3b6e8-ca88-4adf-850b-d167527d15d3-kube-api-access-lvrdk\") pod \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.432143 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-utilities\") pod \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\" (UID: \"dcb3b6e8-ca88-4adf-850b-d167527d15d3\") " Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.433948 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-utilities" (OuterVolumeSpecName: "utilities") pod "dcb3b6e8-ca88-4adf-850b-d167527d15d3" (UID: "dcb3b6e8-ca88-4adf-850b-d167527d15d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.469554 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcb3b6e8-ca88-4adf-850b-d167527d15d3-kube-api-access-lvrdk" (OuterVolumeSpecName: "kube-api-access-lvrdk") pod "dcb3b6e8-ca88-4adf-850b-d167527d15d3" (UID: "dcb3b6e8-ca88-4adf-850b-d167527d15d3"). InnerVolumeSpecName "kube-api-access-lvrdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.487585 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcb3b6e8-ca88-4adf-850b-d167527d15d3" (UID: "dcb3b6e8-ca88-4adf-850b-d167527d15d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.534613 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.534658 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvrdk\" (UniqueName: \"kubernetes.io/projected/dcb3b6e8-ca88-4adf-850b-d167527d15d3-kube-api-access-lvrdk\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.534670 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb3b6e8-ca88-4adf-850b-d167527d15d3-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.908759 4775 generic.go:334] "Generic (PLEG): container finished" podID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerID="7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8" exitCode=0 Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.908825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x9r6" event={"ID":"dcb3b6e8-ca88-4adf-850b-d167527d15d3","Type":"ContainerDied","Data":"7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8"} Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.908852 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8x9r6" event={"ID":"dcb3b6e8-ca88-4adf-850b-d167527d15d3","Type":"ContainerDied","Data":"00f3f36a40004d08a274363f69630267f36788bd8fdfc6fd5a98d5b4f9aa959e"} Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.908857 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8x9r6" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.908882 4775 scope.go:117] "RemoveContainer" containerID="7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.943313 4775 scope.go:117] "RemoveContainer" containerID="13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b" Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.945895 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8x9r6"] Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.955419 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8x9r6"] Nov 26 06:51:38 crc kubenswrapper[4775]: I1126 06:51:38.969086 4775 scope.go:117] "RemoveContainer" containerID="c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.021782 4775 scope.go:117] "RemoveContainer" containerID="7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8" Nov 26 06:51:39 crc kubenswrapper[4775]: E1126 06:51:39.022483 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8\": container with ID starting with 7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8 not found: ID does not exist" containerID="7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.022542 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8"} err="failed to get container status \"7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8\": rpc error: code = NotFound desc = could not find container \"7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8\": container with ID starting with 7a1a773f451b860480e08b49ee5fabb829a94b700e09ec5a3cdd3d89e7d2f4c8 not found: ID does not exist" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.022577 4775 scope.go:117] "RemoveContainer" containerID="13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b" Nov 26 06:51:39 crc kubenswrapper[4775]: E1126 06:51:39.023089 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b\": container with ID starting with 13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b not found: ID does not exist" containerID="13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.023128 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b"} err="failed to get container status \"13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b\": rpc error: code = NotFound desc = could not find container \"13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b\": container with ID starting with 13333cd158bd741f4773bbac7f89be98ab221dff7a1a733d952955851dd57a8b not found: ID does not exist" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.023154 4775 scope.go:117] "RemoveContainer" containerID="c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e" Nov 26 06:51:39 crc kubenswrapper[4775]: E1126 06:51:39.023557 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e\": container with ID starting with c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e not found: ID does not exist" containerID="c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.023592 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e"} err="failed to get container status \"c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e\": rpc error: code = NotFound desc = could not find container \"c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e\": container with ID starting with c22624460d71b0425d41cb4097c1e1c00d60ae5c364b17b3fc0821f47841be5e not found: ID does not exist" Nov 26 06:51:39 crc kubenswrapper[4775]: I1126 06:51:39.351460 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" path="/var/lib/kubelet/pods/dcb3b6e8-ca88-4adf-850b-d167527d15d3/volumes" Nov 26 06:51:53 crc kubenswrapper[4775]: I1126 06:51:53.328852 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:51:53 crc kubenswrapper[4775]: E1126 06:51:53.329871 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:51:58 crc kubenswrapper[4775]: I1126 06:51:58.053079 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hsfmj"] Nov 26 06:51:58 crc kubenswrapper[4775]: I1126 06:51:58.060549 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-mqh7n"] Nov 26 06:51:58 crc kubenswrapper[4775]: I1126 06:51:58.098973 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-mqh7n"] Nov 26 06:51:58 crc kubenswrapper[4775]: I1126 06:51:58.111742 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hsfmj"] Nov 26 06:51:59 crc kubenswrapper[4775]: I1126 06:51:59.349939 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59435a4c-0feb-4d88-8929-e7c0ea22302d" path="/var/lib/kubelet/pods/59435a4c-0feb-4d88-8929-e7c0ea22302d/volumes" Nov 26 06:51:59 crc kubenswrapper[4775]: I1126 06:51:59.352065 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="747654d0-4f51-411e-958b-34f3cd5403eb" path="/var/lib/kubelet/pods/747654d0-4f51-411e-958b-34f3cd5403eb/volumes" Nov 26 06:52:04 crc kubenswrapper[4775]: I1126 06:52:04.045062 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-f8988"] Nov 26 06:52:04 crc kubenswrapper[4775]: I1126 06:52:04.056900 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-f8988"] Nov 26 06:52:04 crc kubenswrapper[4775]: I1126 06:52:04.066291 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-n9tbr"] Nov 26 06:52:04 crc kubenswrapper[4775]: I1126 06:52:04.077583 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-n9tbr"] Nov 26 06:52:04 crc kubenswrapper[4775]: I1126 06:52:04.328169 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:52:04 crc kubenswrapper[4775]: E1126 06:52:04.328826 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:52:05 crc kubenswrapper[4775]: I1126 06:52:05.349667 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25784e47-2e5a-4df0-9aa2-c2a42659cb40" path="/var/lib/kubelet/pods/25784e47-2e5a-4df0-9aa2-c2a42659cb40/volumes" Nov 26 06:52:05 crc kubenswrapper[4775]: I1126 06:52:05.351678 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f163bf-e900-4d6e-b5be-948f927c7b86" path="/var/lib/kubelet/pods/b0f163bf-e900-4d6e-b5be-948f927c7b86/volumes" Nov 26 06:52:09 crc kubenswrapper[4775]: I1126 06:52:09.046545 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-ng8zs"] Nov 26 06:52:09 crc kubenswrapper[4775]: I1126 06:52:09.068943 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-ng8zs"] Nov 26 06:52:09 crc kubenswrapper[4775]: I1126 06:52:09.345933 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d6b7cfb-7b59-4f0b-b36c-94dae19e512d" path="/var/lib/kubelet/pods/9d6b7cfb-7b59-4f0b-b36c-94dae19e512d/volumes" Nov 26 06:52:19 crc kubenswrapper[4775]: I1126 06:52:19.328225 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:52:19 crc kubenswrapper[4775]: E1126 06:52:19.329281 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:52:20 crc kubenswrapper[4775]: I1126 06:52:20.815801 4775 scope.go:117] "RemoveContainer" containerID="312f44c48e89b1dc67f8ba575f2c8bd12f5c18407f3569f8b5a705f41a8a3211" Nov 26 06:52:20 crc kubenswrapper[4775]: I1126 06:52:20.879482 4775 scope.go:117] "RemoveContainer" containerID="b774e3ccc4b186bdb6017dfe69399898f2e6696671dc92422bacfa37e4d3fd85" Nov 26 06:52:20 crc kubenswrapper[4775]: I1126 06:52:20.943012 4775 scope.go:117] "RemoveContainer" containerID="e253f7dee08463f3f7bac7c6e2789f784f447114cc5764ee8b230022d8fca446" Nov 26 06:52:20 crc kubenswrapper[4775]: I1126 06:52:20.967967 4775 scope.go:117] "RemoveContainer" containerID="e5afd9d609d560b067e22b56f0325151bf9412988dfc9ca695c325edba363d09" Nov 26 06:52:21 crc kubenswrapper[4775]: I1126 06:52:21.022292 4775 scope.go:117] "RemoveContainer" containerID="b13fda92c3c8312d7063282c76057773adc40ebd3ea6936dc6aebb2f25e040ec" Nov 26 06:52:22 crc kubenswrapper[4775]: I1126 06:52:22.487982 4775 generic.go:334] "Generic (PLEG): container finished" podID="7b826be9-41b3-4715-b7cb-5b8bc10e35a9" containerID="9210c36c4f40dcc257089d611c183e36e0c0ce11de30877b5939098a710c932c" exitCode=0 Nov 26 06:52:22 crc kubenswrapper[4775]: I1126 06:52:22.488068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" event={"ID":"7b826be9-41b3-4715-b7cb-5b8bc10e35a9","Type":"ContainerDied","Data":"9210c36c4f40dcc257089d611c183e36e0c0ce11de30877b5939098a710c932c"} Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.064690 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.171895 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-inventory\") pod \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.172075 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-ssh-key\") pod \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.172458 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcs6q\" (UniqueName: \"kubernetes.io/projected/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-kube-api-access-bcs6q\") pod \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\" (UID: \"7b826be9-41b3-4715-b7cb-5b8bc10e35a9\") " Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.177883 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-kube-api-access-bcs6q" (OuterVolumeSpecName: "kube-api-access-bcs6q") pod "7b826be9-41b3-4715-b7cb-5b8bc10e35a9" (UID: "7b826be9-41b3-4715-b7cb-5b8bc10e35a9"). InnerVolumeSpecName "kube-api-access-bcs6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.212946 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b826be9-41b3-4715-b7cb-5b8bc10e35a9" (UID: "7b826be9-41b3-4715-b7cb-5b8bc10e35a9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.233274 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-inventory" (OuterVolumeSpecName: "inventory") pod "7b826be9-41b3-4715-b7cb-5b8bc10e35a9" (UID: "7b826be9-41b3-4715-b7cb-5b8bc10e35a9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.275335 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcs6q\" (UniqueName: \"kubernetes.io/projected/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-kube-api-access-bcs6q\") on node \"crc\" DevicePath \"\"" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.275370 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.275382 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b826be9-41b3-4715-b7cb-5b8bc10e35a9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.511080 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" event={"ID":"7b826be9-41b3-4715-b7cb-5b8bc10e35a9","Type":"ContainerDied","Data":"b010a1a3d26be69697e5782780318fc68e7de7d11de2852392aa3ecfbc5b09f8"} Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.511135 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kp72k" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.511141 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b010a1a3d26be69697e5782780318fc68e7de7d11de2852392aa3ecfbc5b09f8" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.625475 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq"] Nov 26 06:52:24 crc kubenswrapper[4775]: E1126 06:52:24.625982 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="extract-content" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.626005 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="extract-content" Nov 26 06:52:24 crc kubenswrapper[4775]: E1126 06:52:24.626041 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="extract-utilities" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.626050 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="extract-utilities" Nov 26 06:52:24 crc kubenswrapper[4775]: E1126 06:52:24.626062 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="registry-server" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.626070 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="registry-server" Nov 26 06:52:24 crc kubenswrapper[4775]: E1126 06:52:24.626082 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b826be9-41b3-4715-b7cb-5b8bc10e35a9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.626098 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b826be9-41b3-4715-b7cb-5b8bc10e35a9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.626310 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b826be9-41b3-4715-b7cb-5b8bc10e35a9" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.626356 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcb3b6e8-ca88-4adf-850b-d167527d15d3" containerName="registry-server" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.627125 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.629125 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.629497 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.629655 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.633135 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.635274 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq"] Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.683248 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.683374 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.683404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdx5p\" (UniqueName: \"kubernetes.io/projected/3d9ee8c3-9727-4454-8df5-5ecce45daf45-kube-api-access-bdx5p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.785250 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.785417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.785477 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdx5p\" (UniqueName: \"kubernetes.io/projected/3d9ee8c3-9727-4454-8df5-5ecce45daf45-kube-api-access-bdx5p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.790362 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.790673 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.806648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdx5p\" (UniqueName: \"kubernetes.io/projected/3d9ee8c3-9727-4454-8df5-5ecce45daf45-kube-api-access-bdx5p\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-z74zq\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:24 crc kubenswrapper[4775]: I1126 06:52:24.966111 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:25 crc kubenswrapper[4775]: I1126 06:52:25.602365 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq"] Nov 26 06:52:25 crc kubenswrapper[4775]: W1126 06:52:25.606828 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d9ee8c3_9727_4454_8df5_5ecce45daf45.slice/crio-00ae6d428e3844c87e4eaec6acad3a20d1a2f3e8525562a43a05e61ac718ce4f WatchSource:0}: Error finding container 00ae6d428e3844c87e4eaec6acad3a20d1a2f3e8525562a43a05e61ac718ce4f: Status 404 returned error can't find the container with id 00ae6d428e3844c87e4eaec6acad3a20d1a2f3e8525562a43a05e61ac718ce4f Nov 26 06:52:26 crc kubenswrapper[4775]: I1126 06:52:26.538856 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" event={"ID":"3d9ee8c3-9727-4454-8df5-5ecce45daf45","Type":"ContainerStarted","Data":"d44b450329cb2672f9e84bb544a2b2934a4bb12576eb8cef13a91a5cf7c281d7"} Nov 26 06:52:26 crc kubenswrapper[4775]: I1126 06:52:26.538938 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" event={"ID":"3d9ee8c3-9727-4454-8df5-5ecce45daf45","Type":"ContainerStarted","Data":"00ae6d428e3844c87e4eaec6acad3a20d1a2f3e8525562a43a05e61ac718ce4f"} Nov 26 06:52:26 crc kubenswrapper[4775]: I1126 06:52:26.569346 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" podStartSLOduration=2.092689378 podStartE2EDuration="2.569317061s" podCreationTimestamp="2025-11-26 06:52:24 +0000 UTC" firstStartedPulling="2025-11-26 06:52:25.609367168 +0000 UTC m=+1628.970671140" lastFinishedPulling="2025-11-26 06:52:26.085994871 +0000 UTC m=+1629.447298823" observedRunningTime="2025-11-26 06:52:26.564622671 +0000 UTC m=+1629.925926643" watchObservedRunningTime="2025-11-26 06:52:26.569317061 +0000 UTC m=+1629.930621053" Nov 26 06:52:31 crc kubenswrapper[4775]: I1126 06:52:31.328895 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:52:31 crc kubenswrapper[4775]: E1126 06:52:31.329704 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:52:31 crc kubenswrapper[4775]: I1126 06:52:31.591352 4775 generic.go:334] "Generic (PLEG): container finished" podID="3d9ee8c3-9727-4454-8df5-5ecce45daf45" containerID="d44b450329cb2672f9e84bb544a2b2934a4bb12576eb8cef13a91a5cf7c281d7" exitCode=0 Nov 26 06:52:31 crc kubenswrapper[4775]: I1126 06:52:31.591431 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" event={"ID":"3d9ee8c3-9727-4454-8df5-5ecce45daf45","Type":"ContainerDied","Data":"d44b450329cb2672f9e84bb544a2b2934a4bb12576eb8cef13a91a5cf7c281d7"} Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.095412 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.154518 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdx5p\" (UniqueName: \"kubernetes.io/projected/3d9ee8c3-9727-4454-8df5-5ecce45daf45-kube-api-access-bdx5p\") pod \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.161669 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9ee8c3-9727-4454-8df5-5ecce45daf45-kube-api-access-bdx5p" (OuterVolumeSpecName: "kube-api-access-bdx5p") pod "3d9ee8c3-9727-4454-8df5-5ecce45daf45" (UID: "3d9ee8c3-9727-4454-8df5-5ecce45daf45"). InnerVolumeSpecName "kube-api-access-bdx5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.255820 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-ssh-key\") pod \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.255990 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-inventory\") pod \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\" (UID: \"3d9ee8c3-9727-4454-8df5-5ecce45daf45\") " Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.256410 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdx5p\" (UniqueName: \"kubernetes.io/projected/3d9ee8c3-9727-4454-8df5-5ecce45daf45-kube-api-access-bdx5p\") on node \"crc\" DevicePath \"\"" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.280187 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-inventory" (OuterVolumeSpecName: "inventory") pod "3d9ee8c3-9727-4454-8df5-5ecce45daf45" (UID: "3d9ee8c3-9727-4454-8df5-5ecce45daf45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.289918 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3d9ee8c3-9727-4454-8df5-5ecce45daf45" (UID: "3d9ee8c3-9727-4454-8df5-5ecce45daf45"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.358345 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.358381 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d9ee8c3-9727-4454-8df5-5ecce45daf45-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.653404 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" event={"ID":"3d9ee8c3-9727-4454-8df5-5ecce45daf45","Type":"ContainerDied","Data":"00ae6d428e3844c87e4eaec6acad3a20d1a2f3e8525562a43a05e61ac718ce4f"} Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.653479 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00ae6d428e3844c87e4eaec6acad3a20d1a2f3e8525562a43a05e61ac718ce4f" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.653646 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-z74zq" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.719867 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl"] Nov 26 06:52:33 crc kubenswrapper[4775]: E1126 06:52:33.720459 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9ee8c3-9727-4454-8df5-5ecce45daf45" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.720490 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9ee8c3-9727-4454-8df5-5ecce45daf45" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.720799 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9ee8c3-9727-4454-8df5-5ecce45daf45" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.721621 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.725036 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.725291 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.726663 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.732678 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.734741 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl"] Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.871820 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swhmg\" (UniqueName: \"kubernetes.io/projected/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-kube-api-access-swhmg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.871879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.872011 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.973479 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.973731 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swhmg\" (UniqueName: \"kubernetes.io/projected/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-kube-api-access-swhmg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.973774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.979038 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.979269 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:33 crc kubenswrapper[4775]: I1126 06:52:33.998064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swhmg\" (UniqueName: \"kubernetes.io/projected/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-kube-api-access-swhmg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dndkl\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:34 crc kubenswrapper[4775]: I1126 06:52:34.084621 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:52:34 crc kubenswrapper[4775]: I1126 06:52:34.397349 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl"] Nov 26 06:52:34 crc kubenswrapper[4775]: I1126 06:52:34.667599 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" event={"ID":"22cadc11-2c9d-49f4-abf7-2e6fa3f69905","Type":"ContainerStarted","Data":"f09482616eb212268a8b88541db2a7098e8785748b3089790c3a5c8ba771e8a2"} Nov 26 06:52:35 crc kubenswrapper[4775]: I1126 06:52:35.682571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" event={"ID":"22cadc11-2c9d-49f4-abf7-2e6fa3f69905","Type":"ContainerStarted","Data":"539452cee8c595d449d30b030fdcd0a9d5b2dfee6399bf9ce33d6b99a306d076"} Nov 26 06:52:35 crc kubenswrapper[4775]: I1126 06:52:35.710381 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" podStartSLOduration=2.227210807 podStartE2EDuration="2.710353412s" podCreationTimestamp="2025-11-26 06:52:33 +0000 UTC" firstStartedPulling="2025-11-26 06:52:34.403223215 +0000 UTC m=+1637.764527167" lastFinishedPulling="2025-11-26 06:52:34.88636581 +0000 UTC m=+1638.247669772" observedRunningTime="2025-11-26 06:52:35.707627996 +0000 UTC m=+1639.068932008" watchObservedRunningTime="2025-11-26 06:52:35.710353412 +0000 UTC m=+1639.071657404" Nov 26 06:52:42 crc kubenswrapper[4775]: I1126 06:52:42.327653 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:52:42 crc kubenswrapper[4775]: E1126 06:52:42.328454 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:52:49 crc kubenswrapper[4775]: I1126 06:52:49.072136 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-35a8-account-create-update-tnxdk"] Nov 26 06:52:49 crc kubenswrapper[4775]: I1126 06:52:49.088578 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-v6zbv"] Nov 26 06:52:49 crc kubenswrapper[4775]: I1126 06:52:49.100846 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-v6zbv"] Nov 26 06:52:49 crc kubenswrapper[4775]: I1126 06:52:49.110599 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-35a8-account-create-update-tnxdk"] Nov 26 06:52:49 crc kubenswrapper[4775]: I1126 06:52:49.347998 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="331d4e96-545d-45e7-b78a-5e42bbc2bdae" path="/var/lib/kubelet/pods/331d4e96-545d-45e7-b78a-5e42bbc2bdae/volumes" Nov 26 06:52:49 crc kubenswrapper[4775]: I1126 06:52:49.349173 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b546f7e2-903a-46dc-8b36-d407c66a86dc" path="/var/lib/kubelet/pods/b546f7e2-903a-46dc-8b36-d407c66a86dc/volumes" Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.031389 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-29ee-account-create-update-f556s"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.042196 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-rktpq"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.052902 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-rktpq"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.063216 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-29ee-account-create-update-f556s"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.070776 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-nqxrj"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.079329 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8bc2-account-create-update-kqltf"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.087902 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-nqxrj"] Nov 26 06:52:50 crc kubenswrapper[4775]: I1126 06:52:50.103423 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8bc2-account-create-update-kqltf"] Nov 26 06:52:51 crc kubenswrapper[4775]: I1126 06:52:51.344908 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="150566f8-cc6b-4ba8-b690-1962b971c9f3" path="/var/lib/kubelet/pods/150566f8-cc6b-4ba8-b690-1962b971c9f3/volumes" Nov 26 06:52:51 crc kubenswrapper[4775]: I1126 06:52:51.347483 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cd7b78e-faa9-4232-a65c-798522553729" path="/var/lib/kubelet/pods/4cd7b78e-faa9-4232-a65c-798522553729/volumes" Nov 26 06:52:51 crc kubenswrapper[4775]: I1126 06:52:51.348908 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a2950b4-e9c2-45f3-9cdc-4bace1b0388b" path="/var/lib/kubelet/pods/5a2950b4-e9c2-45f3-9cdc-4bace1b0388b/volumes" Nov 26 06:52:51 crc kubenswrapper[4775]: I1126 06:52:51.350237 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8" path="/var/lib/kubelet/pods/d4f2f23d-7c87-4a4c-9a9a-aa3786ae98a8/volumes" Nov 26 06:52:57 crc kubenswrapper[4775]: I1126 06:52:57.334114 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:52:57 crc kubenswrapper[4775]: E1126 06:52:57.335030 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:53:12 crc kubenswrapper[4775]: I1126 06:53:12.328807 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:53:12 crc kubenswrapper[4775]: E1126 06:53:12.329941 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:53:16 crc kubenswrapper[4775]: I1126 06:53:16.129089 4775 generic.go:334] "Generic (PLEG): container finished" podID="22cadc11-2c9d-49f4-abf7-2e6fa3f69905" containerID="539452cee8c595d449d30b030fdcd0a9d5b2dfee6399bf9ce33d6b99a306d076" exitCode=0 Nov 26 06:53:16 crc kubenswrapper[4775]: I1126 06:53:16.129205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" event={"ID":"22cadc11-2c9d-49f4-abf7-2e6fa3f69905","Type":"ContainerDied","Data":"539452cee8c595d449d30b030fdcd0a9d5b2dfee6399bf9ce33d6b99a306d076"} Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.585136 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.701614 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-ssh-key\") pod \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.701872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-inventory\") pod \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.702012 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swhmg\" (UniqueName: \"kubernetes.io/projected/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-kube-api-access-swhmg\") pod \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\" (UID: \"22cadc11-2c9d-49f4-abf7-2e6fa3f69905\") " Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.718161 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-kube-api-access-swhmg" (OuterVolumeSpecName: "kube-api-access-swhmg") pod "22cadc11-2c9d-49f4-abf7-2e6fa3f69905" (UID: "22cadc11-2c9d-49f4-abf7-2e6fa3f69905"). InnerVolumeSpecName "kube-api-access-swhmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.753335 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22cadc11-2c9d-49f4-abf7-2e6fa3f69905" (UID: "22cadc11-2c9d-49f4-abf7-2e6fa3f69905"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.770517 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-inventory" (OuterVolumeSpecName: "inventory") pod "22cadc11-2c9d-49f4-abf7-2e6fa3f69905" (UID: "22cadc11-2c9d-49f4-abf7-2e6fa3f69905"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.803845 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swhmg\" (UniqueName: \"kubernetes.io/projected/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-kube-api-access-swhmg\") on node \"crc\" DevicePath \"\"" Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.803883 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:53:17 crc kubenswrapper[4775]: I1126 06:53:17.803897 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22cadc11-2c9d-49f4-abf7-2e6fa3f69905-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.157429 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" event={"ID":"22cadc11-2c9d-49f4-abf7-2e6fa3f69905","Type":"ContainerDied","Data":"f09482616eb212268a8b88541db2a7098e8785748b3089790c3a5c8ba771e8a2"} Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.157495 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f09482616eb212268a8b88541db2a7098e8785748b3089790c3a5c8ba771e8a2" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.157562 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dndkl" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.261305 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g"] Nov 26 06:53:18 crc kubenswrapper[4775]: E1126 06:53:18.261750 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22cadc11-2c9d-49f4-abf7-2e6fa3f69905" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.261772 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="22cadc11-2c9d-49f4-abf7-2e6fa3f69905" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.261979 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="22cadc11-2c9d-49f4-abf7-2e6fa3f69905" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.262574 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.265469 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.266369 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.267013 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.267607 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.284231 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g"] Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.415300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.415345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcf5b\" (UniqueName: \"kubernetes.io/projected/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-kube-api-access-zcf5b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.415682 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.517853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.517939 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcf5b\" (UniqueName: \"kubernetes.io/projected/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-kube-api-access-zcf5b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.518040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.523697 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.534918 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.535227 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcf5b\" (UniqueName: \"kubernetes.io/projected/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-kube-api-access-zcf5b\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:18 crc kubenswrapper[4775]: I1126 06:53:18.604425 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:53:19 crc kubenswrapper[4775]: I1126 06:53:19.046455 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bdd6t"] Nov 26 06:53:19 crc kubenswrapper[4775]: I1126 06:53:19.047399 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bdd6t"] Nov 26 06:53:19 crc kubenswrapper[4775]: I1126 06:53:19.186106 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g"] Nov 26 06:53:19 crc kubenswrapper[4775]: I1126 06:53:19.337716 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1982e34b-5fe7-4993-88dc-418d7786d001" path="/var/lib/kubelet/pods/1982e34b-5fe7-4993-88dc-418d7786d001/volumes" Nov 26 06:53:20 crc kubenswrapper[4775]: I1126 06:53:20.182318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" event={"ID":"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f","Type":"ContainerStarted","Data":"bb055d1930d60edd3880e7e87bcaaf1242e4273eecfaa09abc3272a4dc065feb"} Nov 26 06:53:20 crc kubenswrapper[4775]: I1126 06:53:20.182637 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" event={"ID":"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f","Type":"ContainerStarted","Data":"f739c7d019485706d189613037186549f0102819aa02f66875e0e4bbd0a3666d"} Nov 26 06:53:20 crc kubenswrapper[4775]: I1126 06:53:20.200158 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" podStartSLOduration=1.7145405390000001 podStartE2EDuration="2.200138252s" podCreationTimestamp="2025-11-26 06:53:18 +0000 UTC" firstStartedPulling="2025-11-26 06:53:19.192416963 +0000 UTC m=+1682.553720915" lastFinishedPulling="2025-11-26 06:53:19.678014636 +0000 UTC m=+1683.039318628" observedRunningTime="2025-11-26 06:53:20.198554738 +0000 UTC m=+1683.559858690" watchObservedRunningTime="2025-11-26 06:53:20.200138252 +0000 UTC m=+1683.561442224" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.206398 4775 scope.go:117] "RemoveContainer" containerID="887a3ba4b667f45b7ccd86cf9e4a1e98844e5fc6b934cdb72c21b630df9cae36" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.237959 4775 scope.go:117] "RemoveContainer" containerID="be932c6ad24b9c6d450bd9d0ed52ecafef3127d7bc1d1cc703f0b7f3a6c5ca9b" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.287776 4775 scope.go:117] "RemoveContainer" containerID="1d07ddf20f2967d4dd924b1542743dfc7014c9e95ea705ec23cd638c520995d2" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.325164 4775 scope.go:117] "RemoveContainer" containerID="8573f5ed0b5ab91cf747fc16180f4048b561736f01118aa510a0e58d8db359db" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.369278 4775 scope.go:117] "RemoveContainer" containerID="c90cceebc5643f4620e78dac89d9b6efaf35537fe59fd1f1bbb82f1144e5d1fb" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.427708 4775 scope.go:117] "RemoveContainer" containerID="309a0197f97657300cf97c1cbd375dbd0b23b1f8bf011ece5daa15a89ca7a48a" Nov 26 06:53:21 crc kubenswrapper[4775]: I1126 06:53:21.445766 4775 scope.go:117] "RemoveContainer" containerID="1494c44d9e6c127d6fd7a4c3bb522cfeb9b3129474bbcdbe74e837e7b119498b" Nov 26 06:53:23 crc kubenswrapper[4775]: I1126 06:53:23.328564 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:53:23 crc kubenswrapper[4775]: E1126 06:53:23.329208 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:53:34 crc kubenswrapper[4775]: I1126 06:53:34.348708 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:53:34 crc kubenswrapper[4775]: E1126 06:53:34.349895 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:53:41 crc kubenswrapper[4775]: I1126 06:53:41.037327 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xg4dj"] Nov 26 06:53:41 crc kubenswrapper[4775]: I1126 06:53:41.046787 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xg4dj"] Nov 26 06:53:41 crc kubenswrapper[4775]: I1126 06:53:41.341524 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59a2dfdf-8626-45d9-a37b-91272ecbc570" path="/var/lib/kubelet/pods/59a2dfdf-8626-45d9-a37b-91272ecbc570/volumes" Nov 26 06:53:42 crc kubenswrapper[4775]: I1126 06:53:42.044419 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6prq8"] Nov 26 06:53:42 crc kubenswrapper[4775]: I1126 06:53:42.055599 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6prq8"] Nov 26 06:53:43 crc kubenswrapper[4775]: I1126 06:53:43.350160 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c567945-b393-4d26-8916-cb24e2c38e07" path="/var/lib/kubelet/pods/5c567945-b393-4d26-8916-cb24e2c38e07/volumes" Nov 26 06:53:49 crc kubenswrapper[4775]: I1126 06:53:49.327528 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:53:49 crc kubenswrapper[4775]: E1126 06:53:49.328114 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:54:02 crc kubenswrapper[4775]: I1126 06:54:02.328166 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:54:02 crc kubenswrapper[4775]: E1126 06:54:02.328821 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:54:15 crc kubenswrapper[4775]: I1126 06:54:15.328918 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:54:15 crc kubenswrapper[4775]: E1126 06:54:15.329971 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:54:16 crc kubenswrapper[4775]: I1126 06:54:16.785820 4775 generic.go:334] "Generic (PLEG): container finished" podID="c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" containerID="bb055d1930d60edd3880e7e87bcaaf1242e4273eecfaa09abc3272a4dc065feb" exitCode=0 Nov 26 06:54:16 crc kubenswrapper[4775]: I1126 06:54:16.785936 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" event={"ID":"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f","Type":"ContainerDied","Data":"bb055d1930d60edd3880e7e87bcaaf1242e4273eecfaa09abc3272a4dc065feb"} Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.255672 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.453339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-inventory\") pod \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.453823 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcf5b\" (UniqueName: \"kubernetes.io/projected/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-kube-api-access-zcf5b\") pod \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.453925 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-ssh-key\") pod \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\" (UID: \"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f\") " Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.471791 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-kube-api-access-zcf5b" (OuterVolumeSpecName: "kube-api-access-zcf5b") pod "c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" (UID: "c6dbfaeb-21ee-4017-aab7-6d1219d7e10f"). InnerVolumeSpecName "kube-api-access-zcf5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.483442 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" (UID: "c6dbfaeb-21ee-4017-aab7-6d1219d7e10f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.505152 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-inventory" (OuterVolumeSpecName: "inventory") pod "c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" (UID: "c6dbfaeb-21ee-4017-aab7-6d1219d7e10f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.556635 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcf5b\" (UniqueName: \"kubernetes.io/projected/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-kube-api-access-zcf5b\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.557028 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.557240 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6dbfaeb-21ee-4017-aab7-6d1219d7e10f-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.808388 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" event={"ID":"c6dbfaeb-21ee-4017-aab7-6d1219d7e10f","Type":"ContainerDied","Data":"f739c7d019485706d189613037186549f0102819aa02f66875e0e4bbd0a3666d"} Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.808444 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f739c7d019485706d189613037186549f0102819aa02f66875e0e4bbd0a3666d" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.808447 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.913587 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ghkc9"] Nov 26 06:54:18 crc kubenswrapper[4775]: E1126 06:54:18.914117 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.914141 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.914397 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6dbfaeb-21ee-4017-aab7-6d1219d7e10f" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.915181 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.919567 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.919813 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.919889 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.920134 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:54:18 crc kubenswrapper[4775]: I1126 06:54:18.933225 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ghkc9"] Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.068547 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.068685 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.068778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rjbp\" (UniqueName: \"kubernetes.io/projected/7e9b9292-e385-4016-8355-2b95cf0e85af-kube-api-access-9rjbp\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.170631 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.170727 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.170783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rjbp\" (UniqueName: \"kubernetes.io/projected/7e9b9292-e385-4016-8355-2b95cf0e85af-kube-api-access-9rjbp\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.175644 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.176263 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.198135 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rjbp\" (UniqueName: \"kubernetes.io/projected/7e9b9292-e385-4016-8355-2b95cf0e85af-kube-api-access-9rjbp\") pod \"ssh-known-hosts-edpm-deployment-ghkc9\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.235354 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.767378 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ghkc9"] Nov 26 06:54:19 crc kubenswrapper[4775]: W1126 06:54:19.769171 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e9b9292_e385_4016_8355_2b95cf0e85af.slice/crio-21b60d20234bc08895298dc702a8d84ccefba76ebe77f56750c4c5a3df5958ab WatchSource:0}: Error finding container 21b60d20234bc08895298dc702a8d84ccefba76ebe77f56750c4c5a3df5958ab: Status 404 returned error can't find the container with id 21b60d20234bc08895298dc702a8d84ccefba76ebe77f56750c4c5a3df5958ab Nov 26 06:54:19 crc kubenswrapper[4775]: I1126 06:54:19.817623 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" event={"ID":"7e9b9292-e385-4016-8355-2b95cf0e85af","Type":"ContainerStarted","Data":"21b60d20234bc08895298dc702a8d84ccefba76ebe77f56750c4c5a3df5958ab"} Nov 26 06:54:20 crc kubenswrapper[4775]: I1126 06:54:20.837677 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" event={"ID":"7e9b9292-e385-4016-8355-2b95cf0e85af","Type":"ContainerStarted","Data":"2abb844c1f6c338ce1de4ecae03a41a1d7e7271e02bda0ee538983c471919f94"} Nov 26 06:54:21 crc kubenswrapper[4775]: I1126 06:54:21.602248 4775 scope.go:117] "RemoveContainer" containerID="48a33ff119297c23aba3d9c9db8e7030b84892f7ee4b25bd6d00be5f8e8829b2" Nov 26 06:54:21 crc kubenswrapper[4775]: I1126 06:54:21.672555 4775 scope.go:117] "RemoveContainer" containerID="5cf4f039200a4d89cb81f8faf7b4ffb773a750ebdd1501de701e7d5fb6c34da9" Nov 26 06:54:25 crc kubenswrapper[4775]: I1126 06:54:25.058545 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" podStartSLOduration=6.438804285 podStartE2EDuration="7.058506268s" podCreationTimestamp="2025-11-26 06:54:18 +0000 UTC" firstStartedPulling="2025-11-26 06:54:19.771483128 +0000 UTC m=+1743.132787080" lastFinishedPulling="2025-11-26 06:54:20.391185041 +0000 UTC m=+1743.752489063" observedRunningTime="2025-11-26 06:54:20.866785064 +0000 UTC m=+1744.228089026" watchObservedRunningTime="2025-11-26 06:54:25.058506268 +0000 UTC m=+1748.419810290" Nov 26 06:54:25 crc kubenswrapper[4775]: I1126 06:54:25.068260 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-96s25"] Nov 26 06:54:25 crc kubenswrapper[4775]: I1126 06:54:25.081763 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-96s25"] Nov 26 06:54:25 crc kubenswrapper[4775]: I1126 06:54:25.348008 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ba23799-9d71-495a-a716-56dfca1804a2" path="/var/lib/kubelet/pods/8ba23799-9d71-495a-a716-56dfca1804a2/volumes" Nov 26 06:54:28 crc kubenswrapper[4775]: I1126 06:54:28.921246 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e9b9292-e385-4016-8355-2b95cf0e85af" containerID="2abb844c1f6c338ce1de4ecae03a41a1d7e7271e02bda0ee538983c471919f94" exitCode=0 Nov 26 06:54:28 crc kubenswrapper[4775]: I1126 06:54:28.921362 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" event={"ID":"7e9b9292-e385-4016-8355-2b95cf0e85af","Type":"ContainerDied","Data":"2abb844c1f6c338ce1de4ecae03a41a1d7e7271e02bda0ee538983c471919f94"} Nov 26 06:54:29 crc kubenswrapper[4775]: I1126 06:54:29.328196 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:54:29 crc kubenswrapper[4775]: E1126 06:54:29.328924 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.410416 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.543529 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rjbp\" (UniqueName: \"kubernetes.io/projected/7e9b9292-e385-4016-8355-2b95cf0e85af-kube-api-access-9rjbp\") pod \"7e9b9292-e385-4016-8355-2b95cf0e85af\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.543746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-inventory-0\") pod \"7e9b9292-e385-4016-8355-2b95cf0e85af\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.543832 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-ssh-key-openstack-edpm-ipam\") pod \"7e9b9292-e385-4016-8355-2b95cf0e85af\" (UID: \"7e9b9292-e385-4016-8355-2b95cf0e85af\") " Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.549940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9b9292-e385-4016-8355-2b95cf0e85af-kube-api-access-9rjbp" (OuterVolumeSpecName: "kube-api-access-9rjbp") pod "7e9b9292-e385-4016-8355-2b95cf0e85af" (UID: "7e9b9292-e385-4016-8355-2b95cf0e85af"). InnerVolumeSpecName "kube-api-access-9rjbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.576997 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7e9b9292-e385-4016-8355-2b95cf0e85af" (UID: "7e9b9292-e385-4016-8355-2b95cf0e85af"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.586841 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7e9b9292-e385-4016-8355-2b95cf0e85af" (UID: "7e9b9292-e385-4016-8355-2b95cf0e85af"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.647328 4775 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.647445 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e9b9292-e385-4016-8355-2b95cf0e85af-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.647464 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rjbp\" (UniqueName: \"kubernetes.io/projected/7e9b9292-e385-4016-8355-2b95cf0e85af-kube-api-access-9rjbp\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.939877 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" event={"ID":"7e9b9292-e385-4016-8355-2b95cf0e85af","Type":"ContainerDied","Data":"21b60d20234bc08895298dc702a8d84ccefba76ebe77f56750c4c5a3df5958ab"} Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.939918 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21b60d20234bc08895298dc702a8d84ccefba76ebe77f56750c4c5a3df5958ab" Nov 26 06:54:30 crc kubenswrapper[4775]: I1126 06:54:30.939978 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ghkc9" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.043140 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf"] Nov 26 06:54:31 crc kubenswrapper[4775]: E1126 06:54:31.043572 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9b9292-e385-4016-8355-2b95cf0e85af" containerName="ssh-known-hosts-edpm-deployment" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.043591 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9b9292-e385-4016-8355-2b95cf0e85af" containerName="ssh-known-hosts-edpm-deployment" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.043779 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9b9292-e385-4016-8355-2b95cf0e85af" containerName="ssh-known-hosts-edpm-deployment" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.044362 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.046199 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.046285 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.046811 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.047514 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.063541 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf"] Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.155501 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.155571 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8vhf\" (UniqueName: \"kubernetes.io/projected/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-kube-api-access-p8vhf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.155704 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.257684 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.257813 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8vhf\" (UniqueName: \"kubernetes.io/projected/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-kube-api-access-p8vhf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.257933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.263791 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.264325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.274968 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8vhf\" (UniqueName: \"kubernetes.io/projected/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-kube-api-access-p8vhf\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-lljnf\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.360440 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:31 crc kubenswrapper[4775]: I1126 06:54:31.960387 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf"] Nov 26 06:54:32 crc kubenswrapper[4775]: I1126 06:54:32.956039 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" event={"ID":"6233c3cd-e8bc-4dd1-b039-c950e119ce4e","Type":"ContainerStarted","Data":"38b5e2827f9cb3f4f04e1528cfd7c9be1b36e6ea8918b42f5cac9441f9746354"} Nov 26 06:54:32 crc kubenswrapper[4775]: I1126 06:54:32.956473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" event={"ID":"6233c3cd-e8bc-4dd1-b039-c950e119ce4e","Type":"ContainerStarted","Data":"0b542dafee2d18764404e4167ad9d2c2140e27d2061707904b8b8ace19d7c4ff"} Nov 26 06:54:32 crc kubenswrapper[4775]: I1126 06:54:32.978103 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" podStartSLOduration=1.5192900379999998 podStartE2EDuration="1.97808404s" podCreationTimestamp="2025-11-26 06:54:31 +0000 UTC" firstStartedPulling="2025-11-26 06:54:31.968888607 +0000 UTC m=+1755.330192559" lastFinishedPulling="2025-11-26 06:54:32.427682599 +0000 UTC m=+1755.788986561" observedRunningTime="2025-11-26 06:54:32.976747096 +0000 UTC m=+1756.338051058" watchObservedRunningTime="2025-11-26 06:54:32.97808404 +0000 UTC m=+1756.339387992" Nov 26 06:54:42 crc kubenswrapper[4775]: I1126 06:54:42.056779 4775 generic.go:334] "Generic (PLEG): container finished" podID="6233c3cd-e8bc-4dd1-b039-c950e119ce4e" containerID="38b5e2827f9cb3f4f04e1528cfd7c9be1b36e6ea8918b42f5cac9441f9746354" exitCode=0 Nov 26 06:54:42 crc kubenswrapper[4775]: I1126 06:54:42.056891 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" event={"ID":"6233c3cd-e8bc-4dd1-b039-c950e119ce4e","Type":"ContainerDied","Data":"38b5e2827f9cb3f4f04e1528cfd7c9be1b36e6ea8918b42f5cac9441f9746354"} Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.327917 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:54:43 crc kubenswrapper[4775]: E1126 06:54:43.328497 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.649292 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.657706 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-ssh-key\") pod \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.657810 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8vhf\" (UniqueName: \"kubernetes.io/projected/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-kube-api-access-p8vhf\") pod \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.657840 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-inventory\") pod \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\" (UID: \"6233c3cd-e8bc-4dd1-b039-c950e119ce4e\") " Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.672758 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-kube-api-access-p8vhf" (OuterVolumeSpecName: "kube-api-access-p8vhf") pod "6233c3cd-e8bc-4dd1-b039-c950e119ce4e" (UID: "6233c3cd-e8bc-4dd1-b039-c950e119ce4e"). InnerVolumeSpecName "kube-api-access-p8vhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.697405 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6233c3cd-e8bc-4dd1-b039-c950e119ce4e" (UID: "6233c3cd-e8bc-4dd1-b039-c950e119ce4e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.705311 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-inventory" (OuterVolumeSpecName: "inventory") pod "6233c3cd-e8bc-4dd1-b039-c950e119ce4e" (UID: "6233c3cd-e8bc-4dd1-b039-c950e119ce4e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.760620 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.760674 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8vhf\" (UniqueName: \"kubernetes.io/projected/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-kube-api-access-p8vhf\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:43 crc kubenswrapper[4775]: I1126 06:54:43.760687 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6233c3cd-e8bc-4dd1-b039-c950e119ce4e-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.081216 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" event={"ID":"6233c3cd-e8bc-4dd1-b039-c950e119ce4e","Type":"ContainerDied","Data":"0b542dafee2d18764404e4167ad9d2c2140e27d2061707904b8b8ace19d7c4ff"} Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.081574 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b542dafee2d18764404e4167ad9d2c2140e27d2061707904b8b8ace19d7c4ff" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.081328 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-lljnf" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.171430 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl"] Nov 26 06:54:44 crc kubenswrapper[4775]: E1126 06:54:44.171929 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6233c3cd-e8bc-4dd1-b039-c950e119ce4e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.171952 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6233c3cd-e8bc-4dd1-b039-c950e119ce4e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.172196 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6233c3cd-e8bc-4dd1-b039-c950e119ce4e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.173011 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.175207 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.175424 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.175573 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.176480 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.184625 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl"] Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.268882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.268923 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.268990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw28x\" (UniqueName: \"kubernetes.io/projected/efef5423-9419-4ff8-b603-96ffaa7f4dcd-kube-api-access-bw28x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.371118 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.371181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.371314 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw28x\" (UniqueName: \"kubernetes.io/projected/efef5423-9419-4ff8-b603-96ffaa7f4dcd-kube-api-access-bw28x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.378450 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.378673 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.405180 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw28x\" (UniqueName: \"kubernetes.io/projected/efef5423-9419-4ff8-b603-96ffaa7f4dcd-kube-api-access-bw28x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:44 crc kubenswrapper[4775]: I1126 06:54:44.489032 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:45 crc kubenswrapper[4775]: I1126 06:54:45.038170 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl"] Nov 26 06:54:45 crc kubenswrapper[4775]: I1126 06:54:45.108023 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" event={"ID":"efef5423-9419-4ff8-b603-96ffaa7f4dcd","Type":"ContainerStarted","Data":"be5ac72b6af033d37b9769405aba0843a2e4fbe8bc6e94268d722dad7e7e601b"} Nov 26 06:54:47 crc kubenswrapper[4775]: I1126 06:54:47.129345 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" event={"ID":"efef5423-9419-4ff8-b603-96ffaa7f4dcd","Type":"ContainerStarted","Data":"3ca3420aa7071a600033b6bc33c3b49fc2f4f16f8d51fca42200918e9632ec91"} Nov 26 06:54:47 crc kubenswrapper[4775]: I1126 06:54:47.153381 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" podStartSLOduration=1.7230725850000002 podStartE2EDuration="3.153361656s" podCreationTimestamp="2025-11-26 06:54:44 +0000 UTC" firstStartedPulling="2025-11-26 06:54:45.039629215 +0000 UTC m=+1768.400933167" lastFinishedPulling="2025-11-26 06:54:46.469918276 +0000 UTC m=+1769.831222238" observedRunningTime="2025-11-26 06:54:47.147816732 +0000 UTC m=+1770.509120684" watchObservedRunningTime="2025-11-26 06:54:47.153361656 +0000 UTC m=+1770.514665608" Nov 26 06:54:54 crc kubenswrapper[4775]: I1126 06:54:54.327954 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:54:55 crc kubenswrapper[4775]: I1126 06:54:55.218234 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"89e5f5a059bf9891c4182805221206beffdff0cf5fe2bf0f7b18a435561c8c2c"} Nov 26 06:54:57 crc kubenswrapper[4775]: I1126 06:54:57.239450 4775 generic.go:334] "Generic (PLEG): container finished" podID="efef5423-9419-4ff8-b603-96ffaa7f4dcd" containerID="3ca3420aa7071a600033b6bc33c3b49fc2f4f16f8d51fca42200918e9632ec91" exitCode=0 Nov 26 06:54:57 crc kubenswrapper[4775]: I1126 06:54:57.239983 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" event={"ID":"efef5423-9419-4ff8-b603-96ffaa7f4dcd","Type":"ContainerDied","Data":"3ca3420aa7071a600033b6bc33c3b49fc2f4f16f8d51fca42200918e9632ec91"} Nov 26 06:54:58 crc kubenswrapper[4775]: I1126 06:54:58.828541 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:58 crc kubenswrapper[4775]: I1126 06:54:58.978639 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-inventory\") pod \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " Nov 26 06:54:58 crc kubenswrapper[4775]: I1126 06:54:58.978943 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-ssh-key\") pod \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " Nov 26 06:54:58 crc kubenswrapper[4775]: I1126 06:54:58.979065 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw28x\" (UniqueName: \"kubernetes.io/projected/efef5423-9419-4ff8-b603-96ffaa7f4dcd-kube-api-access-bw28x\") pod \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\" (UID: \"efef5423-9419-4ff8-b603-96ffaa7f4dcd\") " Nov 26 06:54:58 crc kubenswrapper[4775]: I1126 06:54:58.991054 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efef5423-9419-4ff8-b603-96ffaa7f4dcd-kube-api-access-bw28x" (OuterVolumeSpecName: "kube-api-access-bw28x") pod "efef5423-9419-4ff8-b603-96ffaa7f4dcd" (UID: "efef5423-9419-4ff8-b603-96ffaa7f4dcd"). InnerVolumeSpecName "kube-api-access-bw28x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.031499 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-inventory" (OuterVolumeSpecName: "inventory") pod "efef5423-9419-4ff8-b603-96ffaa7f4dcd" (UID: "efef5423-9419-4ff8-b603-96ffaa7f4dcd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.039261 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "efef5423-9419-4ff8-b603-96ffaa7f4dcd" (UID: "efef5423-9419-4ff8-b603-96ffaa7f4dcd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.081816 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.081867 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw28x\" (UniqueName: \"kubernetes.io/projected/efef5423-9419-4ff8-b603-96ffaa7f4dcd-kube-api-access-bw28x\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.081889 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/efef5423-9419-4ff8-b603-96ffaa7f4dcd-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.266662 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" event={"ID":"efef5423-9419-4ff8-b603-96ffaa7f4dcd","Type":"ContainerDied","Data":"be5ac72b6af033d37b9769405aba0843a2e4fbe8bc6e94268d722dad7e7e601b"} Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.266706 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be5ac72b6af033d37b9769405aba0843a2e4fbe8bc6e94268d722dad7e7e601b" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.266830 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.367355 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx"] Nov 26 06:54:59 crc kubenswrapper[4775]: E1126 06:54:59.367818 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efef5423-9419-4ff8-b603-96ffaa7f4dcd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.367840 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="efef5423-9419-4ff8-b603-96ffaa7f4dcd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.368056 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="efef5423-9419-4ff8-b603-96ffaa7f4dcd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.368841 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.371352 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.372094 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.372385 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.373118 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.373546 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.374121 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.375277 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.376174 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.386132 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx"] Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502398 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502538 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502597 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502636 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502746 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbmhl\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-kube-api-access-sbmhl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.502810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503134 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503292 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503444 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503535 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.503622 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605087 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605146 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbmhl\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-kube-api-access-sbmhl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605174 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605241 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605269 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605354 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605379 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.605439 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.610855 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.610976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.611522 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.612204 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.612266 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.612391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.613806 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.614251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.614536 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.614622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.615051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.616000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.620674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.630553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbmhl\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-kube-api-access-sbmhl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-svnzx\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:54:59 crc kubenswrapper[4775]: I1126 06:54:59.692420 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:55:00 crc kubenswrapper[4775]: I1126 06:55:00.024207 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx"] Nov 26 06:55:00 crc kubenswrapper[4775]: I1126 06:55:00.276235 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" event={"ID":"0cad897c-5d2b-496c-8e45-e5061ca93a6a","Type":"ContainerStarted","Data":"cc41f6580aaff26c7a1691eadfa1699ca6c1ec1afd1534b51c6b50ec5142701a"} Nov 26 06:55:01 crc kubenswrapper[4775]: I1126 06:55:01.285027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" event={"ID":"0cad897c-5d2b-496c-8e45-e5061ca93a6a","Type":"ContainerStarted","Data":"c1388f232a1f3bd97cd65359ab57e86b56c9aaf1fead71afa2e2011856493207"} Nov 26 06:55:01 crc kubenswrapper[4775]: I1126 06:55:01.301435 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" podStartSLOduration=1.879679916 podStartE2EDuration="2.301416593s" podCreationTimestamp="2025-11-26 06:54:59 +0000 UTC" firstStartedPulling="2025-11-26 06:55:00.034038957 +0000 UTC m=+1783.395342929" lastFinishedPulling="2025-11-26 06:55:00.455775634 +0000 UTC m=+1783.817079606" observedRunningTime="2025-11-26 06:55:01.299052361 +0000 UTC m=+1784.660356313" watchObservedRunningTime="2025-11-26 06:55:01.301416593 +0000 UTC m=+1784.662720545" Nov 26 06:55:21 crc kubenswrapper[4775]: I1126 06:55:21.816373 4775 scope.go:117] "RemoveContainer" containerID="f140ad268db422358cf37f051ef0d48fa16a5f6a8ef3d211110790f4df383a28" Nov 26 06:55:43 crc kubenswrapper[4775]: I1126 06:55:43.752070 4775 generic.go:334] "Generic (PLEG): container finished" podID="0cad897c-5d2b-496c-8e45-e5061ca93a6a" containerID="c1388f232a1f3bd97cd65359ab57e86b56c9aaf1fead71afa2e2011856493207" exitCode=0 Nov 26 06:55:43 crc kubenswrapper[4775]: I1126 06:55:43.752605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" event={"ID":"0cad897c-5d2b-496c-8e45-e5061ca93a6a","Type":"ContainerDied","Data":"c1388f232a1f3bd97cd65359ab57e86b56c9aaf1fead71afa2e2011856493207"} Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.223573 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.329469 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.329705 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ssh-key\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.329974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-inventory\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.330047 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-neutron-metadata-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.330680 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbmhl\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-kube-api-access-sbmhl\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.330745 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.330812 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-bootstrap-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.330850 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.330890 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-repo-setup-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.331028 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ovn-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.331071 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.331102 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-nova-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.331180 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-telemetry-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.331215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-libvirt-combined-ca-bundle\") pod \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\" (UID: \"0cad897c-5d2b-496c-8e45-e5061ca93a6a\") " Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.336272 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.336434 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.336773 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.337034 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.337169 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.337179 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-kube-api-access-sbmhl" (OuterVolumeSpecName: "kube-api-access-sbmhl") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "kube-api-access-sbmhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.337367 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.339947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.341431 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.341683 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.341821 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.347000 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.363468 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-inventory" (OuterVolumeSpecName: "inventory") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.376883 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0cad897c-5d2b-496c-8e45-e5061ca93a6a" (UID: "0cad897c-5d2b-496c-8e45-e5061ca93a6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433512 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433551 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433566 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbmhl\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-kube-api-access-sbmhl\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433580 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433594 4775 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433607 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433620 4775 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433635 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433647 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433659 4775 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433673 4775 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433684 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433744 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/0cad897c-5d2b-496c-8e45-e5061ca93a6a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.433757 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0cad897c-5d2b-496c-8e45-e5061ca93a6a-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.770532 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" event={"ID":"0cad897c-5d2b-496c-8e45-e5061ca93a6a","Type":"ContainerDied","Data":"cc41f6580aaff26c7a1691eadfa1699ca6c1ec1afd1534b51c6b50ec5142701a"} Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.770885 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc41f6580aaff26c7a1691eadfa1699ca6c1ec1afd1534b51c6b50ec5142701a" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.770951 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-svnzx" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.901480 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7"] Nov 26 06:55:45 crc kubenswrapper[4775]: E1126 06:55:45.901959 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cad897c-5d2b-496c-8e45-e5061ca93a6a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.902003 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cad897c-5d2b-496c-8e45-e5061ca93a6a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.902541 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cad897c-5d2b-496c-8e45-e5061ca93a6a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.903404 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.906228 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.906520 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.906648 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.906788 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.908124 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 26 06:55:45 crc kubenswrapper[4775]: I1126 06:55:45.913928 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7"] Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.043587 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnc45\" (UniqueName: \"kubernetes.io/projected/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-kube-api-access-xnc45\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.043983 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.044311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.044548 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.044794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.146793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.147130 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.147278 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnc45\" (UniqueName: \"kubernetes.io/projected/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-kube-api-access-xnc45\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.147376 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.147503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.148121 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.151693 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.151957 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.153865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.168264 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnc45\" (UniqueName: \"kubernetes.io/projected/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-kube-api-access-xnc45\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gfrt7\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.235157 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:55:46 crc kubenswrapper[4775]: I1126 06:55:46.806985 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7"] Nov 26 06:55:47 crc kubenswrapper[4775]: I1126 06:55:47.797463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" event={"ID":"62bfa377-ebb1-48d4-82b6-eacd0ac68c47","Type":"ContainerStarted","Data":"088d90ba18eed02b2db6b8b1b9ce1b02846b8c0b9a446750cab96428b7a2faf9"} Nov 26 06:55:47 crc kubenswrapper[4775]: I1126 06:55:47.797814 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" event={"ID":"62bfa377-ebb1-48d4-82b6-eacd0ac68c47","Type":"ContainerStarted","Data":"bc911a634c111dcfbd8b0bf6b13c74a5c5cb832fb023b8fb3355e834a4139f59"} Nov 26 06:55:47 crc kubenswrapper[4775]: I1126 06:55:47.825187 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" podStartSLOduration=2.139628483 podStartE2EDuration="2.825164306s" podCreationTimestamp="2025-11-26 06:55:45 +0000 UTC" firstStartedPulling="2025-11-26 06:55:46.817216184 +0000 UTC m=+1830.178520136" lastFinishedPulling="2025-11-26 06:55:47.502751997 +0000 UTC m=+1830.864055959" observedRunningTime="2025-11-26 06:55:47.815021065 +0000 UTC m=+1831.176325087" watchObservedRunningTime="2025-11-26 06:55:47.825164306 +0000 UTC m=+1831.186468288" Nov 26 06:56:54 crc kubenswrapper[4775]: I1126 06:56:54.465473 4775 generic.go:334] "Generic (PLEG): container finished" podID="62bfa377-ebb1-48d4-82b6-eacd0ac68c47" containerID="088d90ba18eed02b2db6b8b1b9ce1b02846b8c0b9a446750cab96428b7a2faf9" exitCode=0 Nov 26 06:56:54 crc kubenswrapper[4775]: I1126 06:56:54.465621 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" event={"ID":"62bfa377-ebb1-48d4-82b6-eacd0ac68c47","Type":"ContainerDied","Data":"088d90ba18eed02b2db6b8b1b9ce1b02846b8c0b9a446750cab96428b7a2faf9"} Nov 26 06:56:55 crc kubenswrapper[4775]: I1126 06:56:55.962792 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.053851 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ssh-key\") pod \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.053928 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-inventory\") pod \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.053965 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnc45\" (UniqueName: \"kubernetes.io/projected/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-kube-api-access-xnc45\") pod \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.054013 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovn-combined-ca-bundle\") pod \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.054117 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovncontroller-config-0\") pod \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\" (UID: \"62bfa377-ebb1-48d4-82b6-eacd0ac68c47\") " Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.060482 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "62bfa377-ebb1-48d4-82b6-eacd0ac68c47" (UID: "62bfa377-ebb1-48d4-82b6-eacd0ac68c47"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.063034 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-kube-api-access-xnc45" (OuterVolumeSpecName: "kube-api-access-xnc45") pod "62bfa377-ebb1-48d4-82b6-eacd0ac68c47" (UID: "62bfa377-ebb1-48d4-82b6-eacd0ac68c47"). InnerVolumeSpecName "kube-api-access-xnc45". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.088505 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-inventory" (OuterVolumeSpecName: "inventory") pod "62bfa377-ebb1-48d4-82b6-eacd0ac68c47" (UID: "62bfa377-ebb1-48d4-82b6-eacd0ac68c47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.111617 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "62bfa377-ebb1-48d4-82b6-eacd0ac68c47" (UID: "62bfa377-ebb1-48d4-82b6-eacd0ac68c47"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.114165 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "62bfa377-ebb1-48d4-82b6-eacd0ac68c47" (UID: "62bfa377-ebb1-48d4-82b6-eacd0ac68c47"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.156354 4775 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.156849 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.156878 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.156905 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnc45\" (UniqueName: \"kubernetes.io/projected/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-kube-api-access-xnc45\") on node \"crc\" DevicePath \"\"" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.156933 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62bfa377-ebb1-48d4-82b6-eacd0ac68c47-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.489218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" event={"ID":"62bfa377-ebb1-48d4-82b6-eacd0ac68c47","Type":"ContainerDied","Data":"bc911a634c111dcfbd8b0bf6b13c74a5c5cb832fb023b8fb3355e834a4139f59"} Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.489257 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc911a634c111dcfbd8b0bf6b13c74a5c5cb832fb023b8fb3355e834a4139f59" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.489346 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gfrt7" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.599056 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27"] Nov 26 06:56:56 crc kubenswrapper[4775]: E1126 06:56:56.599765 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62bfa377-ebb1-48d4-82b6-eacd0ac68c47" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.599801 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="62bfa377-ebb1-48d4-82b6-eacd0ac68c47" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.600083 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="62bfa377-ebb1-48d4-82b6-eacd0ac68c47" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.601025 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.605431 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.606408 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.606596 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.606858 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.606961 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.608794 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.612905 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27"] Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.667959 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.668028 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.668058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.668118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.668166 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.668192 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns852\" (UniqueName: \"kubernetes.io/projected/51e229f8-0c28-4b0c-8307-a2dc128fa503-kube-api-access-ns852\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.770001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.770143 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.770176 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.770222 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.770262 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.770284 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns852\" (UniqueName: \"kubernetes.io/projected/51e229f8-0c28-4b0c-8307-a2dc128fa503-kube-api-access-ns852\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.775535 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.776052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.776423 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.776884 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.779843 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.800432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns852\" (UniqueName: \"kubernetes.io/projected/51e229f8-0c28-4b0c-8307-a2dc128fa503-kube-api-access-ns852\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:56 crc kubenswrapper[4775]: I1126 06:56:56.920678 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:56:57 crc kubenswrapper[4775]: I1126 06:56:57.536730 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27"] Nov 26 06:56:57 crc kubenswrapper[4775]: W1126 06:56:57.543962 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51e229f8_0c28_4b0c_8307_a2dc128fa503.slice/crio-cf8b0f77c70582635e4ea44785cb5c3ab1d03b799a302e9e486edf6871921435 WatchSource:0}: Error finding container cf8b0f77c70582635e4ea44785cb5c3ab1d03b799a302e9e486edf6871921435: Status 404 returned error can't find the container with id cf8b0f77c70582635e4ea44785cb5c3ab1d03b799a302e9e486edf6871921435 Nov 26 06:56:57 crc kubenswrapper[4775]: I1126 06:56:57.546003 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 06:56:58 crc kubenswrapper[4775]: I1126 06:56:58.506036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" event={"ID":"51e229f8-0c28-4b0c-8307-a2dc128fa503","Type":"ContainerStarted","Data":"cf8b0f77c70582635e4ea44785cb5c3ab1d03b799a302e9e486edf6871921435"} Nov 26 06:56:59 crc kubenswrapper[4775]: I1126 06:56:59.515706 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" event={"ID":"51e229f8-0c28-4b0c-8307-a2dc128fa503","Type":"ContainerStarted","Data":"c4a06e27a20c2117455261f3780981b2b23cae189b319fc4ed2817b81bd38216"} Nov 26 06:56:59 crc kubenswrapper[4775]: I1126 06:56:59.538812 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" podStartSLOduration=2.805962955 podStartE2EDuration="3.538791957s" podCreationTimestamp="2025-11-26 06:56:56 +0000 UTC" firstStartedPulling="2025-11-26 06:56:57.54576509 +0000 UTC m=+1900.907069042" lastFinishedPulling="2025-11-26 06:56:58.278594092 +0000 UTC m=+1901.639898044" observedRunningTime="2025-11-26 06:56:59.534827092 +0000 UTC m=+1902.896131054" watchObservedRunningTime="2025-11-26 06:56:59.538791957 +0000 UTC m=+1902.900095909" Nov 26 06:57:21 crc kubenswrapper[4775]: I1126 06:57:21.420076 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:57:21 crc kubenswrapper[4775]: I1126 06:57:21.420896 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:57:50 crc kubenswrapper[4775]: I1126 06:57:50.036309 4775 generic.go:334] "Generic (PLEG): container finished" podID="51e229f8-0c28-4b0c-8307-a2dc128fa503" containerID="c4a06e27a20c2117455261f3780981b2b23cae189b319fc4ed2817b81bd38216" exitCode=0 Nov 26 06:57:50 crc kubenswrapper[4775]: I1126 06:57:50.036359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" event={"ID":"51e229f8-0c28-4b0c-8307-a2dc128fa503","Type":"ContainerDied","Data":"c4a06e27a20c2117455261f3780981b2b23cae189b319fc4ed2817b81bd38216"} Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.420215 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.420608 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.507999 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.587973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-inventory\") pod \"51e229f8-0c28-4b0c-8307-a2dc128fa503\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.588100 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns852\" (UniqueName: \"kubernetes.io/projected/51e229f8-0c28-4b0c-8307-a2dc128fa503-kube-api-access-ns852\") pod \"51e229f8-0c28-4b0c-8307-a2dc128fa503\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.588184 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-nova-metadata-neutron-config-0\") pod \"51e229f8-0c28-4b0c-8307-a2dc128fa503\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.588242 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-metadata-combined-ca-bundle\") pod \"51e229f8-0c28-4b0c-8307-a2dc128fa503\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.588428 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-ovn-metadata-agent-neutron-config-0\") pod \"51e229f8-0c28-4b0c-8307-a2dc128fa503\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.588481 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-ssh-key\") pod \"51e229f8-0c28-4b0c-8307-a2dc128fa503\" (UID: \"51e229f8-0c28-4b0c-8307-a2dc128fa503\") " Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.595791 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51e229f8-0c28-4b0c-8307-a2dc128fa503-kube-api-access-ns852" (OuterVolumeSpecName: "kube-api-access-ns852") pod "51e229f8-0c28-4b0c-8307-a2dc128fa503" (UID: "51e229f8-0c28-4b0c-8307-a2dc128fa503"). InnerVolumeSpecName "kube-api-access-ns852". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.596397 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "51e229f8-0c28-4b0c-8307-a2dc128fa503" (UID: "51e229f8-0c28-4b0c-8307-a2dc128fa503"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.622199 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-inventory" (OuterVolumeSpecName: "inventory") pod "51e229f8-0c28-4b0c-8307-a2dc128fa503" (UID: "51e229f8-0c28-4b0c-8307-a2dc128fa503"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.624917 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "51e229f8-0c28-4b0c-8307-a2dc128fa503" (UID: "51e229f8-0c28-4b0c-8307-a2dc128fa503"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.626601 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "51e229f8-0c28-4b0c-8307-a2dc128fa503" (UID: "51e229f8-0c28-4b0c-8307-a2dc128fa503"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.633660 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "51e229f8-0c28-4b0c-8307-a2dc128fa503" (UID: "51e229f8-0c28-4b0c-8307-a2dc128fa503"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.691249 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.691298 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.691312 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.691326 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns852\" (UniqueName: \"kubernetes.io/projected/51e229f8-0c28-4b0c-8307-a2dc128fa503-kube-api-access-ns852\") on node \"crc\" DevicePath \"\"" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.691338 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 06:57:51 crc kubenswrapper[4775]: I1126 06:57:51.691351 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51e229f8-0c28-4b0c-8307-a2dc128fa503-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.057686 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" event={"ID":"51e229f8-0c28-4b0c-8307-a2dc128fa503","Type":"ContainerDied","Data":"cf8b0f77c70582635e4ea44785cb5c3ab1d03b799a302e9e486edf6871921435"} Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.057754 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf8b0f77c70582635e4ea44785cb5c3ab1d03b799a302e9e486edf6871921435" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.057767 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.224866 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j"] Nov 26 06:57:52 crc kubenswrapper[4775]: E1126 06:57:52.225269 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51e229f8-0c28-4b0c-8307-a2dc128fa503" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.225288 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="51e229f8-0c28-4b0c-8307-a2dc128fa503" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.225522 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="51e229f8-0c28-4b0c-8307-a2dc128fa503" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.226212 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.228480 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.228678 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.228897 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.229064 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.231848 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.244646 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j"] Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.302511 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.302554 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qt8r\" (UniqueName: \"kubernetes.io/projected/da48dd91-7390-4751-8e73-455fda656b0d-kube-api-access-4qt8r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.302621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.302642 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.302660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.404821 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.405178 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.405858 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.406122 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qt8r\" (UniqueName: \"kubernetes.io/projected/da48dd91-7390-4751-8e73-455fda656b0d-kube-api-access-4qt8r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.407753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.410462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.411088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.411186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.412636 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.424887 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qt8r\" (UniqueName: \"kubernetes.io/projected/da48dd91-7390-4751-8e73-455fda656b0d-kube-api-access-4qt8r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:52 crc kubenswrapper[4775]: I1126 06:57:52.545088 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 06:57:53 crc kubenswrapper[4775]: I1126 06:57:53.066432 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j"] Nov 26 06:57:54 crc kubenswrapper[4775]: I1126 06:57:54.078184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" event={"ID":"da48dd91-7390-4751-8e73-455fda656b0d","Type":"ContainerStarted","Data":"927252a75f8b17468714d2869c14ab105c26d2b2a6aac816af653fb102e9f6e5"} Nov 26 06:57:54 crc kubenswrapper[4775]: I1126 06:57:54.078646 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" event={"ID":"da48dd91-7390-4751-8e73-455fda656b0d","Type":"ContainerStarted","Data":"a64cf7518bd8dd46db4a2919e37f4198b3aad8cdca4d063c75ad64d6c4643e57"} Nov 26 06:57:54 crc kubenswrapper[4775]: I1126 06:57:54.104131 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" podStartSLOduration=1.683607249 podStartE2EDuration="2.104115175s" podCreationTimestamp="2025-11-26 06:57:52 +0000 UTC" firstStartedPulling="2025-11-26 06:57:53.069012206 +0000 UTC m=+1956.430316158" lastFinishedPulling="2025-11-26 06:57:53.489520122 +0000 UTC m=+1956.850824084" observedRunningTime="2025-11-26 06:57:54.101151998 +0000 UTC m=+1957.462455950" watchObservedRunningTime="2025-11-26 06:57:54.104115175 +0000 UTC m=+1957.465419127" Nov 26 06:58:21 crc kubenswrapper[4775]: I1126 06:58:21.419649 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 06:58:21 crc kubenswrapper[4775]: I1126 06:58:21.420286 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 06:58:21 crc kubenswrapper[4775]: I1126 06:58:21.420338 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 06:58:21 crc kubenswrapper[4775]: I1126 06:58:21.421299 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89e5f5a059bf9891c4182805221206beffdff0cf5fe2bf0f7b18a435561c8c2c"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 06:58:21 crc kubenswrapper[4775]: I1126 06:58:21.421354 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://89e5f5a059bf9891c4182805221206beffdff0cf5fe2bf0f7b18a435561c8c2c" gracePeriod=600 Nov 26 06:58:22 crc kubenswrapper[4775]: I1126 06:58:22.356072 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="89e5f5a059bf9891c4182805221206beffdff0cf5fe2bf0f7b18a435561c8c2c" exitCode=0 Nov 26 06:58:22 crc kubenswrapper[4775]: I1126 06:58:22.356158 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"89e5f5a059bf9891c4182805221206beffdff0cf5fe2bf0f7b18a435561c8c2c"} Nov 26 06:58:22 crc kubenswrapper[4775]: I1126 06:58:22.356601 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689"} Nov 26 06:58:22 crc kubenswrapper[4775]: I1126 06:58:22.356624 4775 scope.go:117] "RemoveContainer" containerID="eea10f61bc9a118b79d660ee7057851ebdc78384f9a4d89d50cc530c8fd50e56" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.184930 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nt9ql"] Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.188106 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.210446 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpd7d\" (UniqueName: \"kubernetes.io/projected/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-kube-api-access-dpd7d\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.210616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-utilities\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.210752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-catalog-content\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.205459 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nt9ql"] Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.312043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-utilities\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.312148 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-catalog-content\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.312221 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpd7d\" (UniqueName: \"kubernetes.io/projected/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-kube-api-access-dpd7d\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.312568 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-utilities\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.312659 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-catalog-content\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.331422 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpd7d\" (UniqueName: \"kubernetes.io/projected/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-kube-api-access-dpd7d\") pod \"certified-operators-nt9ql\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:51 crc kubenswrapper[4775]: I1126 06:58:51.516945 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:58:52 crc kubenswrapper[4775]: I1126 06:58:52.013932 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nt9ql"] Nov 26 06:58:52 crc kubenswrapper[4775]: I1126 06:58:52.648162 4775 generic.go:334] "Generic (PLEG): container finished" podID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerID="98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c" exitCode=0 Nov 26 06:58:52 crc kubenswrapper[4775]: I1126 06:58:52.648228 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nt9ql" event={"ID":"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a","Type":"ContainerDied","Data":"98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c"} Nov 26 06:58:52 crc kubenswrapper[4775]: I1126 06:58:52.648268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nt9ql" event={"ID":"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a","Type":"ContainerStarted","Data":"f4f12d6c4dba6a47bbc811ac7a9ffd73d1783e3c4949c98ddc923d7c54fdd265"} Nov 26 06:58:54 crc kubenswrapper[4775]: I1126 06:58:54.673607 4775 generic.go:334] "Generic (PLEG): container finished" podID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerID="3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25" exitCode=0 Nov 26 06:58:54 crc kubenswrapper[4775]: I1126 06:58:54.673681 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nt9ql" event={"ID":"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a","Type":"ContainerDied","Data":"3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25"} Nov 26 06:58:55 crc kubenswrapper[4775]: I1126 06:58:55.690801 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nt9ql" event={"ID":"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a","Type":"ContainerStarted","Data":"e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd"} Nov 26 06:58:55 crc kubenswrapper[4775]: I1126 06:58:55.715834 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nt9ql" podStartSLOduration=2.283014443 podStartE2EDuration="4.71581311s" podCreationTimestamp="2025-11-26 06:58:51 +0000 UTC" firstStartedPulling="2025-11-26 06:58:52.650217102 +0000 UTC m=+2016.011521084" lastFinishedPulling="2025-11-26 06:58:55.083015769 +0000 UTC m=+2018.444319751" observedRunningTime="2025-11-26 06:58:55.714351033 +0000 UTC m=+2019.075655055" watchObservedRunningTime="2025-11-26 06:58:55.71581311 +0000 UTC m=+2019.077117052" Nov 26 06:59:01 crc kubenswrapper[4775]: I1126 06:59:01.517566 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:59:01 crc kubenswrapper[4775]: I1126 06:59:01.518133 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:59:01 crc kubenswrapper[4775]: I1126 06:59:01.576083 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:59:01 crc kubenswrapper[4775]: I1126 06:59:01.824817 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:59:04 crc kubenswrapper[4775]: I1126 06:59:04.529768 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nt9ql"] Nov 26 06:59:04 crc kubenswrapper[4775]: I1126 06:59:04.531174 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nt9ql" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="registry-server" containerID="cri-o://e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd" gracePeriod=2 Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.089944 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.146034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-catalog-content\") pod \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.146315 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-utilities\") pod \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.146381 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpd7d\" (UniqueName: \"kubernetes.io/projected/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-kube-api-access-dpd7d\") pod \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\" (UID: \"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a\") " Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.147362 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-utilities" (OuterVolumeSpecName: "utilities") pod "5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" (UID: "5b56c3c4-0dff-4c49-bbf6-4b7913dd330a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.151948 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-kube-api-access-dpd7d" (OuterVolumeSpecName: "kube-api-access-dpd7d") pod "5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" (UID: "5b56c3c4-0dff-4c49-bbf6-4b7913dd330a"). InnerVolumeSpecName "kube-api-access-dpd7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.193316 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" (UID: "5b56c3c4-0dff-4c49-bbf6-4b7913dd330a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.228667 4775 generic.go:334] "Generic (PLEG): container finished" podID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerID="e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd" exitCode=0 Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.228703 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nt9ql" event={"ID":"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a","Type":"ContainerDied","Data":"e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd"} Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.228753 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nt9ql" event={"ID":"5b56c3c4-0dff-4c49-bbf6-4b7913dd330a","Type":"ContainerDied","Data":"f4f12d6c4dba6a47bbc811ac7a9ffd73d1783e3c4949c98ddc923d7c54fdd265"} Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.228778 4775 scope.go:117] "RemoveContainer" containerID="e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.228795 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nt9ql" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.248406 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.248446 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpd7d\" (UniqueName: \"kubernetes.io/projected/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-kube-api-access-dpd7d\") on node \"crc\" DevicePath \"\"" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.248459 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.250244 4775 scope.go:117] "RemoveContainer" containerID="3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.295279 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nt9ql"] Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.297869 4775 scope.go:117] "RemoveContainer" containerID="98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.311467 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nt9ql"] Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.323209 4775 scope.go:117] "RemoveContainer" containerID="e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd" Nov 26 06:59:05 crc kubenswrapper[4775]: E1126 06:59:05.323679 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd\": container with ID starting with e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd not found: ID does not exist" containerID="e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.323740 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd"} err="failed to get container status \"e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd\": rpc error: code = NotFound desc = could not find container \"e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd\": container with ID starting with e5192e560ddd047fbb63222676689ecd18d177906daefc5c7cadd2babfa476dd not found: ID does not exist" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.323770 4775 scope.go:117] "RemoveContainer" containerID="3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25" Nov 26 06:59:05 crc kubenswrapper[4775]: E1126 06:59:05.324184 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25\": container with ID starting with 3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25 not found: ID does not exist" containerID="3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.324221 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25"} err="failed to get container status \"3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25\": rpc error: code = NotFound desc = could not find container \"3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25\": container with ID starting with 3d7a27625b94f9119a6e85d6acb6592e6538e6135f0b621a0f5e830d564f2e25 not found: ID does not exist" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.324241 4775 scope.go:117] "RemoveContainer" containerID="98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c" Nov 26 06:59:05 crc kubenswrapper[4775]: E1126 06:59:05.324527 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c\": container with ID starting with 98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c not found: ID does not exist" containerID="98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.324557 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c"} err="failed to get container status \"98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c\": rpc error: code = NotFound desc = could not find container \"98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c\": container with ID starting with 98d5f9cda963c836765cf7eec924931c09d725f9a0b0f8d559c4754ebe31cc5c not found: ID does not exist" Nov 26 06:59:05 crc kubenswrapper[4775]: I1126 06:59:05.336746 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" path="/var/lib/kubelet/pods/5b56c3c4-0dff-4c49-bbf6-4b7913dd330a/volumes" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.160969 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9"] Nov 26 07:00:00 crc kubenswrapper[4775]: E1126 07:00:00.161931 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="extract-content" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.161949 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="extract-content" Nov 26 07:00:00 crc kubenswrapper[4775]: E1126 07:00:00.161977 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="extract-utilities" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.161986 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="extract-utilities" Nov 26 07:00:00 crc kubenswrapper[4775]: E1126 07:00:00.162019 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="registry-server" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.162027 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="registry-server" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.162245 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b56c3c4-0dff-4c49-bbf6-4b7913dd330a" containerName="registry-server" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.163067 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.164991 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.170563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9"] Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.171106 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.311594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqmv6\" (UniqueName: \"kubernetes.io/projected/b4364d09-a73b-424d-b08c-271493dbd918-kube-api-access-dqmv6\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.311660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4364d09-a73b-424d-b08c-271493dbd918-secret-volume\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.311790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4364d09-a73b-424d-b08c-271493dbd918-config-volume\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.414263 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqmv6\" (UniqueName: \"kubernetes.io/projected/b4364d09-a73b-424d-b08c-271493dbd918-kube-api-access-dqmv6\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.414312 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4364d09-a73b-424d-b08c-271493dbd918-secret-volume\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.414376 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4364d09-a73b-424d-b08c-271493dbd918-config-volume\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.416304 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4364d09-a73b-424d-b08c-271493dbd918-config-volume\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.424418 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4364d09-a73b-424d-b08c-271493dbd918-secret-volume\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.442908 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqmv6\" (UniqueName: \"kubernetes.io/projected/b4364d09-a73b-424d-b08c-271493dbd918-kube-api-access-dqmv6\") pod \"collect-profiles-29402340-pzcg9\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.494589 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:00 crc kubenswrapper[4775]: I1126 07:00:00.940525 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9"] Nov 26 07:00:01 crc kubenswrapper[4775]: I1126 07:00:01.730190 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4364d09-a73b-424d-b08c-271493dbd918" containerID="4dbf6de87e2b90dcfc59a6f3deed9ca78f28b92029ecb986aa89404a247cb50f" exitCode=0 Nov 26 07:00:01 crc kubenswrapper[4775]: I1126 07:00:01.730351 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" event={"ID":"b4364d09-a73b-424d-b08c-271493dbd918","Type":"ContainerDied","Data":"4dbf6de87e2b90dcfc59a6f3deed9ca78f28b92029ecb986aa89404a247cb50f"} Nov 26 07:00:01 crc kubenswrapper[4775]: I1126 07:00:01.730792 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" event={"ID":"b4364d09-a73b-424d-b08c-271493dbd918","Type":"ContainerStarted","Data":"4c6e5b9075f2f5fbc49c033632f7f2785735291c14bce4202d9e3ea5c92e6a3b"} Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.097304 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.268170 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4364d09-a73b-424d-b08c-271493dbd918-config-volume\") pod \"b4364d09-a73b-424d-b08c-271493dbd918\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.268377 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4364d09-a73b-424d-b08c-271493dbd918-secret-volume\") pod \"b4364d09-a73b-424d-b08c-271493dbd918\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.268433 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqmv6\" (UniqueName: \"kubernetes.io/projected/b4364d09-a73b-424d-b08c-271493dbd918-kube-api-access-dqmv6\") pod \"b4364d09-a73b-424d-b08c-271493dbd918\" (UID: \"b4364d09-a73b-424d-b08c-271493dbd918\") " Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.269094 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4364d09-a73b-424d-b08c-271493dbd918-config-volume" (OuterVolumeSpecName: "config-volume") pod "b4364d09-a73b-424d-b08c-271493dbd918" (UID: "b4364d09-a73b-424d-b08c-271493dbd918"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.274947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4364d09-a73b-424d-b08c-271493dbd918-kube-api-access-dqmv6" (OuterVolumeSpecName: "kube-api-access-dqmv6") pod "b4364d09-a73b-424d-b08c-271493dbd918" (UID: "b4364d09-a73b-424d-b08c-271493dbd918"). InnerVolumeSpecName "kube-api-access-dqmv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.275136 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4364d09-a73b-424d-b08c-271493dbd918-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b4364d09-a73b-424d-b08c-271493dbd918" (UID: "b4364d09-a73b-424d-b08c-271493dbd918"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.371072 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4364d09-a73b-424d-b08c-271493dbd918-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.371122 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4364d09-a73b-424d-b08c-271493dbd918-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.371148 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqmv6\" (UniqueName: \"kubernetes.io/projected/b4364d09-a73b-424d-b08c-271493dbd918-kube-api-access-dqmv6\") on node \"crc\" DevicePath \"\"" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.757208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" event={"ID":"b4364d09-a73b-424d-b08c-271493dbd918","Type":"ContainerDied","Data":"4c6e5b9075f2f5fbc49c033632f7f2785735291c14bce4202d9e3ea5c92e6a3b"} Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.757257 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c6e5b9075f2f5fbc49c033632f7f2785735291c14bce4202d9e3ea5c92e6a3b" Nov 26 07:00:03 crc kubenswrapper[4775]: I1126 07:00:03.757262 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402340-pzcg9" Nov 26 07:00:04 crc kubenswrapper[4775]: I1126 07:00:04.195498 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg"] Nov 26 07:00:04 crc kubenswrapper[4775]: I1126 07:00:04.205939 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402295-bqjbg"] Nov 26 07:00:05 crc kubenswrapper[4775]: I1126 07:00:05.577088 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3" path="/var/lib/kubelet/pods/ef5a7cb0-e9d9-4ec4-8855-da8e3685fbd3/volumes" Nov 26 07:00:21 crc kubenswrapper[4775]: I1126 07:00:21.420108 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:00:21 crc kubenswrapper[4775]: I1126 07:00:21.421814 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:00:21 crc kubenswrapper[4775]: I1126 07:00:21.994047 4775 scope.go:117] "RemoveContainer" containerID="d95be0bed38af2377ee70712b06f96a07c0956b6629e4603ad5aeee39865f917" Nov 26 07:00:51 crc kubenswrapper[4775]: I1126 07:00:51.420146 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:00:51 crc kubenswrapper[4775]: I1126 07:00:51.420660 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.148683 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29402341-lzbf6"] Nov 26 07:01:00 crc kubenswrapper[4775]: E1126 07:01:00.152709 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4364d09-a73b-424d-b08c-271493dbd918" containerName="collect-profiles" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.152746 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4364d09-a73b-424d-b08c-271493dbd918" containerName="collect-profiles" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.153181 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4364d09-a73b-424d-b08c-271493dbd918" containerName="collect-profiles" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.154015 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.163907 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29402341-lzbf6"] Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.298984 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-combined-ca-bundle\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.299179 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-fernet-keys\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.299225 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-config-data\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.299273 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x44pr\" (UniqueName: \"kubernetes.io/projected/55d5f81a-1fab-4bfa-8ca9-115de4fef547-kube-api-access-x44pr\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.401023 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-fernet-keys\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.401092 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-config-data\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.401137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x44pr\" (UniqueName: \"kubernetes.io/projected/55d5f81a-1fab-4bfa-8ca9-115de4fef547-kube-api-access-x44pr\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.401215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-combined-ca-bundle\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.409646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-config-data\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.409692 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-combined-ca-bundle\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.411788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-fernet-keys\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.420200 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x44pr\" (UniqueName: \"kubernetes.io/projected/55d5f81a-1fab-4bfa-8ca9-115de4fef547-kube-api-access-x44pr\") pod \"keystone-cron-29402341-lzbf6\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.479916 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:00 crc kubenswrapper[4775]: I1126 07:01:00.952526 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29402341-lzbf6"] Nov 26 07:01:01 crc kubenswrapper[4775]: I1126 07:01:01.279518 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402341-lzbf6" event={"ID":"55d5f81a-1fab-4bfa-8ca9-115de4fef547","Type":"ContainerStarted","Data":"6e0f5582b08e7675d02302d63e5cb073d6a11ea7fe8e525386b8ce0138f56cfa"} Nov 26 07:01:02 crc kubenswrapper[4775]: I1126 07:01:02.295494 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402341-lzbf6" event={"ID":"55d5f81a-1fab-4bfa-8ca9-115de4fef547","Type":"ContainerStarted","Data":"2f126980cb7eba5d1e4305ada0fa2a17c7e56c6f634775c69ab0d7454462303d"} Nov 26 07:01:02 crc kubenswrapper[4775]: I1126 07:01:02.319303 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29402341-lzbf6" podStartSLOduration=2.319279848 podStartE2EDuration="2.319279848s" podCreationTimestamp="2025-11-26 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 07:01:02.316813035 +0000 UTC m=+2145.678116987" watchObservedRunningTime="2025-11-26 07:01:02.319279848 +0000 UTC m=+2145.680583810" Nov 26 07:01:04 crc kubenswrapper[4775]: I1126 07:01:04.317376 4775 generic.go:334] "Generic (PLEG): container finished" podID="55d5f81a-1fab-4bfa-8ca9-115de4fef547" containerID="2f126980cb7eba5d1e4305ada0fa2a17c7e56c6f634775c69ab0d7454462303d" exitCode=0 Nov 26 07:01:04 crc kubenswrapper[4775]: I1126 07:01:04.317705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402341-lzbf6" event={"ID":"55d5f81a-1fab-4bfa-8ca9-115de4fef547","Type":"ContainerDied","Data":"2f126980cb7eba5d1e4305ada0fa2a17c7e56c6f634775c69ab0d7454462303d"} Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.766218 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.812331 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x44pr\" (UniqueName: \"kubernetes.io/projected/55d5f81a-1fab-4bfa-8ca9-115de4fef547-kube-api-access-x44pr\") pod \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.812467 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-fernet-keys\") pod \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.812490 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-combined-ca-bundle\") pod \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.812555 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-config-data\") pod \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\" (UID: \"55d5f81a-1fab-4bfa-8ca9-115de4fef547\") " Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.820707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d5f81a-1fab-4bfa-8ca9-115de4fef547-kube-api-access-x44pr" (OuterVolumeSpecName: "kube-api-access-x44pr") pod "55d5f81a-1fab-4bfa-8ca9-115de4fef547" (UID: "55d5f81a-1fab-4bfa-8ca9-115de4fef547"). InnerVolumeSpecName "kube-api-access-x44pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.824866 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "55d5f81a-1fab-4bfa-8ca9-115de4fef547" (UID: "55d5f81a-1fab-4bfa-8ca9-115de4fef547"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.853625 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55d5f81a-1fab-4bfa-8ca9-115de4fef547" (UID: "55d5f81a-1fab-4bfa-8ca9-115de4fef547"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.872621 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-config-data" (OuterVolumeSpecName: "config-data") pod "55d5f81a-1fab-4bfa-8ca9-115de4fef547" (UID: "55d5f81a-1fab-4bfa-8ca9-115de4fef547"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.914453 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x44pr\" (UniqueName: \"kubernetes.io/projected/55d5f81a-1fab-4bfa-8ca9-115de4fef547-kube-api-access-x44pr\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.914484 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.914494 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:05 crc kubenswrapper[4775]: I1126 07:01:05.914503 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55d5f81a-1fab-4bfa-8ca9-115de4fef547-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:06 crc kubenswrapper[4775]: I1126 07:01:06.342379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29402341-lzbf6" event={"ID":"55d5f81a-1fab-4bfa-8ca9-115de4fef547","Type":"ContainerDied","Data":"6e0f5582b08e7675d02302d63e5cb073d6a11ea7fe8e525386b8ce0138f56cfa"} Nov 26 07:01:06 crc kubenswrapper[4775]: I1126 07:01:06.342418 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e0f5582b08e7675d02302d63e5cb073d6a11ea7fe8e525386b8ce0138f56cfa" Nov 26 07:01:06 crc kubenswrapper[4775]: I1126 07:01:06.342466 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29402341-lzbf6" Nov 26 07:01:21 crc kubenswrapper[4775]: I1126 07:01:21.420202 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:01:21 crc kubenswrapper[4775]: I1126 07:01:21.420863 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:01:21 crc kubenswrapper[4775]: I1126 07:01:21.420920 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:01:21 crc kubenswrapper[4775]: I1126 07:01:21.421781 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:01:21 crc kubenswrapper[4775]: I1126 07:01:21.421849 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" gracePeriod=600 Nov 26 07:01:21 crc kubenswrapper[4775]: E1126 07:01:21.562948 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:01:22 crc kubenswrapper[4775]: I1126 07:01:22.523212 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" exitCode=0 Nov 26 07:01:22 crc kubenswrapper[4775]: I1126 07:01:22.523268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689"} Nov 26 07:01:22 crc kubenswrapper[4775]: I1126 07:01:22.523567 4775 scope.go:117] "RemoveContainer" containerID="89e5f5a059bf9891c4182805221206beffdff0cf5fe2bf0f7b18a435561c8c2c" Nov 26 07:01:22 crc kubenswrapper[4775]: I1126 07:01:22.524184 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:01:22 crc kubenswrapper[4775]: E1126 07:01:22.524502 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:01:34 crc kubenswrapper[4775]: I1126 07:01:34.327751 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:01:34 crc kubenswrapper[4775]: E1126 07:01:34.328433 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.392338 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dm9k7"] Nov 26 07:01:38 crc kubenswrapper[4775]: E1126 07:01:38.404509 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d5f81a-1fab-4bfa-8ca9-115de4fef547" containerName="keystone-cron" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.404552 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d5f81a-1fab-4bfa-8ca9-115de4fef547" containerName="keystone-cron" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.404945 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d5f81a-1fab-4bfa-8ca9-115de4fef547" containerName="keystone-cron" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.407311 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.420165 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm9k7"] Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.496337 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-catalog-content\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.496573 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rsvv\" (UniqueName: \"kubernetes.io/projected/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-kube-api-access-5rsvv\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.496669 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-utilities\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.598098 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-catalog-content\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.598167 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rsvv\" (UniqueName: \"kubernetes.io/projected/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-kube-api-access-5rsvv\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.598214 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-utilities\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.598700 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-utilities\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.598751 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-catalog-content\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.618194 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rsvv\" (UniqueName: \"kubernetes.io/projected/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-kube-api-access-5rsvv\") pod \"redhat-marketplace-dm9k7\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:38 crc kubenswrapper[4775]: I1126 07:01:38.730388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:39 crc kubenswrapper[4775]: I1126 07:01:39.201892 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm9k7"] Nov 26 07:01:39 crc kubenswrapper[4775]: I1126 07:01:39.672944 4775 generic.go:334] "Generic (PLEG): container finished" podID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerID="ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b" exitCode=0 Nov 26 07:01:39 crc kubenswrapper[4775]: I1126 07:01:39.673035 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm9k7" event={"ID":"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2","Type":"ContainerDied","Data":"ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b"} Nov 26 07:01:39 crc kubenswrapper[4775]: I1126 07:01:39.673326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm9k7" event={"ID":"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2","Type":"ContainerStarted","Data":"765e53b765b340fdbc22d9db539980a7f4d12111693d80d83f9c00d14184151a"} Nov 26 07:01:41 crc kubenswrapper[4775]: I1126 07:01:41.699094 4775 generic.go:334] "Generic (PLEG): container finished" podID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerID="a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74" exitCode=0 Nov 26 07:01:41 crc kubenswrapper[4775]: I1126 07:01:41.699211 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm9k7" event={"ID":"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2","Type":"ContainerDied","Data":"a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74"} Nov 26 07:01:42 crc kubenswrapper[4775]: I1126 07:01:42.712564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm9k7" event={"ID":"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2","Type":"ContainerStarted","Data":"99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38"} Nov 26 07:01:42 crc kubenswrapper[4775]: I1126 07:01:42.732411 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dm9k7" podStartSLOduration=2.300199174 podStartE2EDuration="4.732390787s" podCreationTimestamp="2025-11-26 07:01:38 +0000 UTC" firstStartedPulling="2025-11-26 07:01:39.675315612 +0000 UTC m=+2183.036619564" lastFinishedPulling="2025-11-26 07:01:42.107507225 +0000 UTC m=+2185.468811177" observedRunningTime="2025-11-26 07:01:42.730077337 +0000 UTC m=+2186.091381309" watchObservedRunningTime="2025-11-26 07:01:42.732390787 +0000 UTC m=+2186.093694729" Nov 26 07:01:46 crc kubenswrapper[4775]: I1126 07:01:46.328830 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:01:46 crc kubenswrapper[4775]: E1126 07:01:46.329699 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:01:48 crc kubenswrapper[4775]: I1126 07:01:48.731597 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:48 crc kubenswrapper[4775]: I1126 07:01:48.731967 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:48 crc kubenswrapper[4775]: I1126 07:01:48.793633 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:48 crc kubenswrapper[4775]: I1126 07:01:48.848229 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:49 crc kubenswrapper[4775]: I1126 07:01:49.033608 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm9k7"] Nov 26 07:01:50 crc kubenswrapper[4775]: I1126 07:01:50.802194 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dm9k7" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="registry-server" containerID="cri-o://99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38" gracePeriod=2 Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.260498 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.340549 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-catalog-content\") pod \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.340928 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-utilities\") pod \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.341026 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rsvv\" (UniqueName: \"kubernetes.io/projected/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-kube-api-access-5rsvv\") pod \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\" (UID: \"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2\") " Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.342531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-utilities" (OuterVolumeSpecName: "utilities") pod "627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" (UID: "627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.353043 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-kube-api-access-5rsvv" (OuterVolumeSpecName: "kube-api-access-5rsvv") pod "627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" (UID: "627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2"). InnerVolumeSpecName "kube-api-access-5rsvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.359568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" (UID: "627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.443544 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.443588 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rsvv\" (UniqueName: \"kubernetes.io/projected/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-kube-api-access-5rsvv\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.443599 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.817454 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dm9k7" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.817566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm9k7" event={"ID":"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2","Type":"ContainerDied","Data":"99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38"} Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.818993 4775 scope.go:117] "RemoveContainer" containerID="99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.817342 4775 generic.go:334] "Generic (PLEG): container finished" podID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerID="99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38" exitCode=0 Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.823927 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dm9k7" event={"ID":"627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2","Type":"ContainerDied","Data":"765e53b765b340fdbc22d9db539980a7f4d12111693d80d83f9c00d14184151a"} Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.843619 4775 scope.go:117] "RemoveContainer" containerID="a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.861431 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm9k7"] Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.870527 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dm9k7"] Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.883783 4775 scope.go:117] "RemoveContainer" containerID="ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.928918 4775 scope.go:117] "RemoveContainer" containerID="99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38" Nov 26 07:01:51 crc kubenswrapper[4775]: E1126 07:01:51.929498 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38\": container with ID starting with 99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38 not found: ID does not exist" containerID="99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.929585 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38"} err="failed to get container status \"99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38\": rpc error: code = NotFound desc = could not find container \"99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38\": container with ID starting with 99d4d4527aad3d6366fa841ada1d745d9c2db7cc71d539d919bf2d894a18bb38 not found: ID does not exist" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.929656 4775 scope.go:117] "RemoveContainer" containerID="a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74" Nov 26 07:01:51 crc kubenswrapper[4775]: E1126 07:01:51.930108 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74\": container with ID starting with a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74 not found: ID does not exist" containerID="a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.930184 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74"} err="failed to get container status \"a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74\": rpc error: code = NotFound desc = could not find container \"a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74\": container with ID starting with a4cff69b50418bb58cedee62193cf377078d1d3ccc34f419b3e72ed0e6838d74 not found: ID does not exist" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.930255 4775 scope.go:117] "RemoveContainer" containerID="ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b" Nov 26 07:01:51 crc kubenswrapper[4775]: E1126 07:01:51.930665 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b\": container with ID starting with ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b not found: ID does not exist" containerID="ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b" Nov 26 07:01:51 crc kubenswrapper[4775]: I1126 07:01:51.930744 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b"} err="failed to get container status \"ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b\": rpc error: code = NotFound desc = could not find container \"ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b\": container with ID starting with ffe2c0e7f18fcae4d60838684abcd7d3af086baf9fbbed42dbb345d5f889d89b not found: ID does not exist" Nov 26 07:01:53 crc kubenswrapper[4775]: I1126 07:01:53.343262 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" path="/var/lib/kubelet/pods/627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2/volumes" Nov 26 07:02:01 crc kubenswrapper[4775]: I1126 07:02:01.328054 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:02:01 crc kubenswrapper[4775]: E1126 07:02:01.328917 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.943400 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fjb9j"] Nov 26 07:02:12 crc kubenswrapper[4775]: E1126 07:02:12.944380 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="registry-server" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.944395 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="registry-server" Nov 26 07:02:12 crc kubenswrapper[4775]: E1126 07:02:12.944418 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="extract-content" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.944426 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="extract-content" Nov 26 07:02:12 crc kubenswrapper[4775]: E1126 07:02:12.944439 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="extract-utilities" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.944448 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="extract-utilities" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.944743 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="627dfbb3-45ad-4ed7-bddf-d9d78c6a20b2" containerName="registry-server" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.946515 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:12 crc kubenswrapper[4775]: I1126 07:02:12.960700 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fjb9j"] Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.004546 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-catalog-content\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.004734 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnj29\" (UniqueName: \"kubernetes.io/projected/7993caa1-b677-44ba-a755-26f4c130fab2-kube-api-access-bnj29\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.004784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-utilities\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.106293 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-catalog-content\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.106834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnj29\" (UniqueName: \"kubernetes.io/projected/7993caa1-b677-44ba-a755-26f4c130fab2-kube-api-access-bnj29\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.106855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-utilities\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.106863 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-catalog-content\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.107149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-utilities\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.128007 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnj29\" (UniqueName: \"kubernetes.io/projected/7993caa1-b677-44ba-a755-26f4c130fab2-kube-api-access-bnj29\") pod \"community-operators-fjb9j\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.306468 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:13 crc kubenswrapper[4775]: I1126 07:02:13.888734 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fjb9j"] Nov 26 07:02:13 crc kubenswrapper[4775]: W1126 07:02:13.900193 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7993caa1_b677_44ba_a755_26f4c130fab2.slice/crio-d63e8a8edc2da266c15124b11de0080665b2b4b2131c25ac7c99ec3698b086fd WatchSource:0}: Error finding container d63e8a8edc2da266c15124b11de0080665b2b4b2131c25ac7c99ec3698b086fd: Status 404 returned error can't find the container with id d63e8a8edc2da266c15124b11de0080665b2b4b2131c25ac7c99ec3698b086fd Nov 26 07:02:14 crc kubenswrapper[4775]: I1126 07:02:14.026161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerStarted","Data":"d63e8a8edc2da266c15124b11de0080665b2b4b2131c25ac7c99ec3698b086fd"} Nov 26 07:02:14 crc kubenswrapper[4775]: I1126 07:02:14.027916 4775 generic.go:334] "Generic (PLEG): container finished" podID="da48dd91-7390-4751-8e73-455fda656b0d" containerID="927252a75f8b17468714d2869c14ab105c26d2b2a6aac816af653fb102e9f6e5" exitCode=0 Nov 26 07:02:14 crc kubenswrapper[4775]: I1126 07:02:14.027963 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" event={"ID":"da48dd91-7390-4751-8e73-455fda656b0d","Type":"ContainerDied","Data":"927252a75f8b17468714d2869c14ab105c26d2b2a6aac816af653fb102e9f6e5"} Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.045634 4775 generic.go:334] "Generic (PLEG): container finished" podID="7993caa1-b677-44ba-a755-26f4c130fab2" containerID="a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4" exitCode=0 Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.045703 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerDied","Data":"a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4"} Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.050026 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.570019 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.672061 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-inventory\") pod \"da48dd91-7390-4751-8e73-455fda656b0d\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.672140 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-ssh-key\") pod \"da48dd91-7390-4751-8e73-455fda656b0d\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.672170 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-combined-ca-bundle\") pod \"da48dd91-7390-4751-8e73-455fda656b0d\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.672603 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-secret-0\") pod \"da48dd91-7390-4751-8e73-455fda656b0d\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.672643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qt8r\" (UniqueName: \"kubernetes.io/projected/da48dd91-7390-4751-8e73-455fda656b0d-kube-api-access-4qt8r\") pod \"da48dd91-7390-4751-8e73-455fda656b0d\" (UID: \"da48dd91-7390-4751-8e73-455fda656b0d\") " Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.678654 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "da48dd91-7390-4751-8e73-455fda656b0d" (UID: "da48dd91-7390-4751-8e73-455fda656b0d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.679643 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da48dd91-7390-4751-8e73-455fda656b0d-kube-api-access-4qt8r" (OuterVolumeSpecName: "kube-api-access-4qt8r") pod "da48dd91-7390-4751-8e73-455fda656b0d" (UID: "da48dd91-7390-4751-8e73-455fda656b0d"). InnerVolumeSpecName "kube-api-access-4qt8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.707418 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "da48dd91-7390-4751-8e73-455fda656b0d" (UID: "da48dd91-7390-4751-8e73-455fda656b0d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.713693 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "da48dd91-7390-4751-8e73-455fda656b0d" (UID: "da48dd91-7390-4751-8e73-455fda656b0d"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.725917 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-inventory" (OuterVolumeSpecName: "inventory") pod "da48dd91-7390-4751-8e73-455fda656b0d" (UID: "da48dd91-7390-4751-8e73-455fda656b0d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.775147 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.775189 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qt8r\" (UniqueName: \"kubernetes.io/projected/da48dd91-7390-4751-8e73-455fda656b0d-kube-api-access-4qt8r\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.775206 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.775216 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:15 crc kubenswrapper[4775]: I1126 07:02:15.775228 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da48dd91-7390-4751-8e73-455fda656b0d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.058855 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" event={"ID":"da48dd91-7390-4751-8e73-455fda656b0d","Type":"ContainerDied","Data":"a64cf7518bd8dd46db4a2919e37f4198b3aad8cdca4d063c75ad64d6c4643e57"} Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.059127 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a64cf7518bd8dd46db4a2919e37f4198b3aad8cdca4d063c75ad64d6c4643e57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.058891 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.064561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerStarted","Data":"c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d"} Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.146190 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57"] Nov 26 07:02:16 crc kubenswrapper[4775]: E1126 07:02:16.146627 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da48dd91-7390-4751-8e73-455fda656b0d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.146649 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="da48dd91-7390-4751-8e73-455fda656b0d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.146919 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="da48dd91-7390-4751-8e73-455fda656b0d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.147624 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.154972 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.154997 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.155091 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.155485 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.155880 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.156005 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.189110 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.192281 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57"] Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284595 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284671 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284836 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s448x\" (UniqueName: \"kubernetes.io/projected/fc8364bf-1ce5-4681-99a5-edc754f142ac-kube-api-access-s448x\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284904 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284965 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.284994 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.285019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.285066 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.327885 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:02:16 crc kubenswrapper[4775]: E1126 07:02:16.328164 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.386459 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.386513 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.386541 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.386927 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.386979 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.387092 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.387162 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.387210 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s448x\" (UniqueName: \"kubernetes.io/projected/fc8364bf-1ce5-4681-99a5-edc754f142ac-kube-api-access-s448x\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.387300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.390914 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.394329 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.395375 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.395614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.396000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.397437 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.398014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.399115 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.410196 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s448x\" (UniqueName: \"kubernetes.io/projected/fc8364bf-1ce5-4681-99a5-edc754f142ac-kube-api-access-s448x\") pod \"nova-edpm-deployment-openstack-edpm-ipam-ldl57\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.465863 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:02:16 crc kubenswrapper[4775]: I1126 07:02:16.982227 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57"] Nov 26 07:02:16 crc kubenswrapper[4775]: W1126 07:02:16.983161 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc8364bf_1ce5_4681_99a5_edc754f142ac.slice/crio-f1b8972c5bbf2d1c670c32acf07ca62a0560784f02334c72d4892c0494126f75 WatchSource:0}: Error finding container f1b8972c5bbf2d1c670c32acf07ca62a0560784f02334c72d4892c0494126f75: Status 404 returned error can't find the container with id f1b8972c5bbf2d1c670c32acf07ca62a0560784f02334c72d4892c0494126f75 Nov 26 07:02:17 crc kubenswrapper[4775]: I1126 07:02:17.075081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" event={"ID":"fc8364bf-1ce5-4681-99a5-edc754f142ac","Type":"ContainerStarted","Data":"f1b8972c5bbf2d1c670c32acf07ca62a0560784f02334c72d4892c0494126f75"} Nov 26 07:02:17 crc kubenswrapper[4775]: I1126 07:02:17.078512 4775 generic.go:334] "Generic (PLEG): container finished" podID="7993caa1-b677-44ba-a755-26f4c130fab2" containerID="c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d" exitCode=0 Nov 26 07:02:17 crc kubenswrapper[4775]: I1126 07:02:17.078553 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerDied","Data":"c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d"} Nov 26 07:02:17 crc kubenswrapper[4775]: I1126 07:02:17.477293 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.087941 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" event={"ID":"fc8364bf-1ce5-4681-99a5-edc754f142ac","Type":"ContainerStarted","Data":"b646f09287d1f055853b6c4b73838886877771df6e728006a2bdff5c626004a3"} Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.091693 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerStarted","Data":"1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2"} Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.107464 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" podStartSLOduration=1.6185415669999998 podStartE2EDuration="2.107438486s" podCreationTimestamp="2025-11-26 07:02:16 +0000 UTC" firstStartedPulling="2025-11-26 07:02:16.9862948 +0000 UTC m=+2220.347598752" lastFinishedPulling="2025-11-26 07:02:17.475191729 +0000 UTC m=+2220.836495671" observedRunningTime="2025-11-26 07:02:18.105337321 +0000 UTC m=+2221.466641273" watchObservedRunningTime="2025-11-26 07:02:18.107438486 +0000 UTC m=+2221.468742428" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.129271 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fjb9j" podStartSLOduration=3.674898307 podStartE2EDuration="6.129253859s" podCreationTimestamp="2025-11-26 07:02:12 +0000 UTC" firstStartedPulling="2025-11-26 07:02:15.048104461 +0000 UTC m=+2218.409408443" lastFinishedPulling="2025-11-26 07:02:17.502460043 +0000 UTC m=+2220.863763995" observedRunningTime="2025-11-26 07:02:18.127907407 +0000 UTC m=+2221.489211399" watchObservedRunningTime="2025-11-26 07:02:18.129253859 +0000 UTC m=+2221.490557811" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.723794 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rnfqx"] Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.729796 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.735381 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnfqx"] Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.836886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmvzv\" (UniqueName: \"kubernetes.io/projected/514b3d70-4c1d-47a4-b700-5c4381ee371d-kube-api-access-lmvzv\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.836954 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-catalog-content\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.837027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-utilities\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.938762 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmvzv\" (UniqueName: \"kubernetes.io/projected/514b3d70-4c1d-47a4-b700-5c4381ee371d-kube-api-access-lmvzv\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.939174 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-catalog-content\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.939437 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-utilities\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.939743 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-catalog-content\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.939853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-utilities\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:18 crc kubenswrapper[4775]: I1126 07:02:18.966826 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmvzv\" (UniqueName: \"kubernetes.io/projected/514b3d70-4c1d-47a4-b700-5c4381ee371d-kube-api-access-lmvzv\") pod \"redhat-operators-rnfqx\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:19 crc kubenswrapper[4775]: I1126 07:02:19.047023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:19 crc kubenswrapper[4775]: I1126 07:02:19.510144 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnfqx"] Nov 26 07:02:19 crc kubenswrapper[4775]: W1126 07:02:19.513993 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod514b3d70_4c1d_47a4_b700_5c4381ee371d.slice/crio-84c750889e687a7abcbf86b051bd1e0111353b817ca145a15cdf4460b069ae05 WatchSource:0}: Error finding container 84c750889e687a7abcbf86b051bd1e0111353b817ca145a15cdf4460b069ae05: Status 404 returned error can't find the container with id 84c750889e687a7abcbf86b051bd1e0111353b817ca145a15cdf4460b069ae05 Nov 26 07:02:20 crc kubenswrapper[4775]: I1126 07:02:20.110236 4775 generic.go:334] "Generic (PLEG): container finished" podID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerID="cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71" exitCode=0 Nov 26 07:02:20 crc kubenswrapper[4775]: I1126 07:02:20.110290 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerDied","Data":"cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71"} Nov 26 07:02:20 crc kubenswrapper[4775]: I1126 07:02:20.110323 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerStarted","Data":"84c750889e687a7abcbf86b051bd1e0111353b817ca145a15cdf4460b069ae05"} Nov 26 07:02:21 crc kubenswrapper[4775]: I1126 07:02:21.119005 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerStarted","Data":"db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd"} Nov 26 07:02:23 crc kubenswrapper[4775]: I1126 07:02:23.307674 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:23 crc kubenswrapper[4775]: I1126 07:02:23.308292 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:23 crc kubenswrapper[4775]: I1126 07:02:23.351218 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:24 crc kubenswrapper[4775]: I1126 07:02:24.154147 4775 generic.go:334] "Generic (PLEG): container finished" podID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerID="db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd" exitCode=0 Nov 26 07:02:24 crc kubenswrapper[4775]: I1126 07:02:24.154213 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerDied","Data":"db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd"} Nov 26 07:02:24 crc kubenswrapper[4775]: I1126 07:02:24.225809 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:24 crc kubenswrapper[4775]: I1126 07:02:24.721025 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fjb9j"] Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.173589 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerStarted","Data":"0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea"} Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.173799 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fjb9j" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="registry-server" containerID="cri-o://1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2" gracePeriod=2 Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.208381 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rnfqx" podStartSLOduration=2.961696789 podStartE2EDuration="8.208365375s" podCreationTimestamp="2025-11-26 07:02:18 +0000 UTC" firstStartedPulling="2025-11-26 07:02:20.112383525 +0000 UTC m=+2223.473687477" lastFinishedPulling="2025-11-26 07:02:25.359052101 +0000 UTC m=+2228.720356063" observedRunningTime="2025-11-26 07:02:26.202787741 +0000 UTC m=+2229.564091703" watchObservedRunningTime="2025-11-26 07:02:26.208365375 +0000 UTC m=+2229.569669327" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.661795 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.704387 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-utilities\") pod \"7993caa1-b677-44ba-a755-26f4c130fab2\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.704567 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnj29\" (UniqueName: \"kubernetes.io/projected/7993caa1-b677-44ba-a755-26f4c130fab2-kube-api-access-bnj29\") pod \"7993caa1-b677-44ba-a755-26f4c130fab2\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.704702 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-catalog-content\") pod \"7993caa1-b677-44ba-a755-26f4c130fab2\" (UID: \"7993caa1-b677-44ba-a755-26f4c130fab2\") " Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.705420 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-utilities" (OuterVolumeSpecName: "utilities") pod "7993caa1-b677-44ba-a755-26f4c130fab2" (UID: "7993caa1-b677-44ba-a755-26f4c130fab2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.705627 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.718821 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7993caa1-b677-44ba-a755-26f4c130fab2-kube-api-access-bnj29" (OuterVolumeSpecName: "kube-api-access-bnj29") pod "7993caa1-b677-44ba-a755-26f4c130fab2" (UID: "7993caa1-b677-44ba-a755-26f4c130fab2"). InnerVolumeSpecName "kube-api-access-bnj29". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.766753 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7993caa1-b677-44ba-a755-26f4c130fab2" (UID: "7993caa1-b677-44ba-a755-26f4c130fab2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.807357 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnj29\" (UniqueName: \"kubernetes.io/projected/7993caa1-b677-44ba-a755-26f4c130fab2-kube-api-access-bnj29\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:26 crc kubenswrapper[4775]: I1126 07:02:26.807599 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7993caa1-b677-44ba-a755-26f4c130fab2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.185909 4775 generic.go:334] "Generic (PLEG): container finished" podID="7993caa1-b677-44ba-a755-26f4c130fab2" containerID="1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2" exitCode=0 Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.185972 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerDied","Data":"1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2"} Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.186033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fjb9j" event={"ID":"7993caa1-b677-44ba-a755-26f4c130fab2","Type":"ContainerDied","Data":"d63e8a8edc2da266c15124b11de0080665b2b4b2131c25ac7c99ec3698b086fd"} Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.186056 4775 scope.go:117] "RemoveContainer" containerID="1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.188266 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fjb9j" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.210586 4775 scope.go:117] "RemoveContainer" containerID="c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.237335 4775 scope.go:117] "RemoveContainer" containerID="a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.258351 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fjb9j"] Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.266980 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fjb9j"] Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.298075 4775 scope.go:117] "RemoveContainer" containerID="1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2" Nov 26 07:02:27 crc kubenswrapper[4775]: E1126 07:02:27.311158 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2\": container with ID starting with 1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2 not found: ID does not exist" containerID="1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.311211 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2"} err="failed to get container status \"1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2\": rpc error: code = NotFound desc = could not find container \"1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2\": container with ID starting with 1c0b45214a84a2f2c88f5db38acb9cd280a34112ebc013d08e94edc6bfbb29a2 not found: ID does not exist" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.311243 4775 scope.go:117] "RemoveContainer" containerID="c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d" Nov 26 07:02:27 crc kubenswrapper[4775]: E1126 07:02:27.311516 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d\": container with ID starting with c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d not found: ID does not exist" containerID="c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.311602 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d"} err="failed to get container status \"c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d\": rpc error: code = NotFound desc = could not find container \"c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d\": container with ID starting with c8d673749bb72f7a204fafd51a273c25f2bb1f158347e8403402c58192b6c29d not found: ID does not exist" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.311674 4775 scope.go:117] "RemoveContainer" containerID="a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4" Nov 26 07:02:27 crc kubenswrapper[4775]: E1126 07:02:27.312299 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4\": container with ID starting with a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4 not found: ID does not exist" containerID="a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.312350 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4"} err="failed to get container status \"a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4\": rpc error: code = NotFound desc = could not find container \"a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4\": container with ID starting with a5704abc94bf4a512ebfb2bece99da5e11c081b4e3661579254c163c99c21ce4 not found: ID does not exist" Nov 26 07:02:27 crc kubenswrapper[4775]: I1126 07:02:27.337529 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" path="/var/lib/kubelet/pods/7993caa1-b677-44ba-a755-26f4c130fab2/volumes" Nov 26 07:02:29 crc kubenswrapper[4775]: I1126 07:02:29.047792 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:29 crc kubenswrapper[4775]: I1126 07:02:29.048187 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:29 crc kubenswrapper[4775]: I1126 07:02:29.328923 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:02:29 crc kubenswrapper[4775]: E1126 07:02:29.329506 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:02:30 crc kubenswrapper[4775]: I1126 07:02:30.112494 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rnfqx" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="registry-server" probeResult="failure" output=< Nov 26 07:02:30 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Nov 26 07:02:30 crc kubenswrapper[4775]: > Nov 26 07:02:39 crc kubenswrapper[4775]: I1126 07:02:39.107058 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:39 crc kubenswrapper[4775]: I1126 07:02:39.158988 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:39 crc kubenswrapper[4775]: I1126 07:02:39.351797 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnfqx"] Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.306141 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rnfqx" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="registry-server" containerID="cri-o://0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea" gracePeriod=2 Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.327866 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:02:40 crc kubenswrapper[4775]: E1126 07:02:40.328383 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.747621 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.799461 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-utilities\") pod \"514b3d70-4c1d-47a4-b700-5c4381ee371d\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.799592 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-catalog-content\") pod \"514b3d70-4c1d-47a4-b700-5c4381ee371d\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.799654 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmvzv\" (UniqueName: \"kubernetes.io/projected/514b3d70-4c1d-47a4-b700-5c4381ee371d-kube-api-access-lmvzv\") pod \"514b3d70-4c1d-47a4-b700-5c4381ee371d\" (UID: \"514b3d70-4c1d-47a4-b700-5c4381ee371d\") " Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.800531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-utilities" (OuterVolumeSpecName: "utilities") pod "514b3d70-4c1d-47a4-b700-5c4381ee371d" (UID: "514b3d70-4c1d-47a4-b700-5c4381ee371d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.806833 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/514b3d70-4c1d-47a4-b700-5c4381ee371d-kube-api-access-lmvzv" (OuterVolumeSpecName: "kube-api-access-lmvzv") pod "514b3d70-4c1d-47a4-b700-5c4381ee371d" (UID: "514b3d70-4c1d-47a4-b700-5c4381ee371d"). InnerVolumeSpecName "kube-api-access-lmvzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.902620 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmvzv\" (UniqueName: \"kubernetes.io/projected/514b3d70-4c1d-47a4-b700-5c4381ee371d-kube-api-access-lmvzv\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.902663 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:40 crc kubenswrapper[4775]: I1126 07:02:40.911560 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "514b3d70-4c1d-47a4-b700-5c4381ee371d" (UID: "514b3d70-4c1d-47a4-b700-5c4381ee371d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.005233 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/514b3d70-4c1d-47a4-b700-5c4381ee371d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.318806 4775 generic.go:334] "Generic (PLEG): container finished" podID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerID="0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea" exitCode=0 Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.318885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerDied","Data":"0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea"} Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.318931 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnfqx" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.318946 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnfqx" event={"ID":"514b3d70-4c1d-47a4-b700-5c4381ee371d","Type":"ContainerDied","Data":"84c750889e687a7abcbf86b051bd1e0111353b817ca145a15cdf4460b069ae05"} Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.318991 4775 scope.go:117] "RemoveContainer" containerID="0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.354326 4775 scope.go:117] "RemoveContainer" containerID="db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.369100 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnfqx"] Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.378385 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rnfqx"] Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.386325 4775 scope.go:117] "RemoveContainer" containerID="cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.427487 4775 scope.go:117] "RemoveContainer" containerID="0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea" Nov 26 07:02:41 crc kubenswrapper[4775]: E1126 07:02:41.428064 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea\": container with ID starting with 0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea not found: ID does not exist" containerID="0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.428134 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea"} err="failed to get container status \"0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea\": rpc error: code = NotFound desc = could not find container \"0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea\": container with ID starting with 0efb55fae073e73c018c583000e6b814007dbdacba0ac2c06135910299c2d2ea not found: ID does not exist" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.428166 4775 scope.go:117] "RemoveContainer" containerID="db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd" Nov 26 07:02:41 crc kubenswrapper[4775]: E1126 07:02:41.428526 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd\": container with ID starting with db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd not found: ID does not exist" containerID="db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.428560 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd"} err="failed to get container status \"db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd\": rpc error: code = NotFound desc = could not find container \"db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd\": container with ID starting with db41705ea2b25d95b638fe068ecb774346ef8e2e5251745c39ca770a25d8edcd not found: ID does not exist" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.428590 4775 scope.go:117] "RemoveContainer" containerID="cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71" Nov 26 07:02:41 crc kubenswrapper[4775]: E1126 07:02:41.428885 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71\": container with ID starting with cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71 not found: ID does not exist" containerID="cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71" Nov 26 07:02:41 crc kubenswrapper[4775]: I1126 07:02:41.428924 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71"} err="failed to get container status \"cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71\": rpc error: code = NotFound desc = could not find container \"cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71\": container with ID starting with cef0f9ff67a39d2ffce977cc044c7ccb3fa990fa5301ec337db843100cf0da71 not found: ID does not exist" Nov 26 07:02:43 crc kubenswrapper[4775]: I1126 07:02:43.350938 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" path="/var/lib/kubelet/pods/514b3d70-4c1d-47a4-b700-5c4381ee371d/volumes" Nov 26 07:02:53 crc kubenswrapper[4775]: I1126 07:02:53.327334 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:02:53 crc kubenswrapper[4775]: E1126 07:02:53.328038 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:03:08 crc kubenswrapper[4775]: I1126 07:03:08.328293 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:03:08 crc kubenswrapper[4775]: E1126 07:03:08.329133 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:03:23 crc kubenswrapper[4775]: I1126 07:03:23.328374 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:03:23 crc kubenswrapper[4775]: E1126 07:03:23.335839 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:03:37 crc kubenswrapper[4775]: I1126 07:03:37.354132 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:03:37 crc kubenswrapper[4775]: E1126 07:03:37.354905 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:03:48 crc kubenswrapper[4775]: I1126 07:03:48.329140 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:03:48 crc kubenswrapper[4775]: E1126 07:03:48.330021 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:04:01 crc kubenswrapper[4775]: I1126 07:04:01.328395 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:04:01 crc kubenswrapper[4775]: E1126 07:04:01.329360 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:04:15 crc kubenswrapper[4775]: I1126 07:04:15.328321 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:04:15 crc kubenswrapper[4775]: E1126 07:04:15.329085 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:04:30 crc kubenswrapper[4775]: I1126 07:04:30.328256 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:04:30 crc kubenswrapper[4775]: E1126 07:04:30.330208 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:04:42 crc kubenswrapper[4775]: I1126 07:04:42.328264 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:04:42 crc kubenswrapper[4775]: E1126 07:04:42.329182 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:04:54 crc kubenswrapper[4775]: I1126 07:04:54.328268 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:04:54 crc kubenswrapper[4775]: E1126 07:04:54.329029 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:05:09 crc kubenswrapper[4775]: I1126 07:05:09.328916 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:05:09 crc kubenswrapper[4775]: E1126 07:05:09.330444 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:05:11 crc kubenswrapper[4775]: I1126 07:05:11.747771 4775 generic.go:334] "Generic (PLEG): container finished" podID="fc8364bf-1ce5-4681-99a5-edc754f142ac" containerID="b646f09287d1f055853b6c4b73838886877771df6e728006a2bdff5c626004a3" exitCode=0 Nov 26 07:05:11 crc kubenswrapper[4775]: I1126 07:05:11.747894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" event={"ID":"fc8364bf-1ce5-4681-99a5-edc754f142ac","Type":"ContainerDied","Data":"b646f09287d1f055853b6c4b73838886877771df6e728006a2bdff5c626004a3"} Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.217643 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379277 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-1\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379341 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-extra-config-0\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379369 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-1\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379477 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s448x\" (UniqueName: \"kubernetes.io/projected/fc8364bf-1ce5-4681-99a5-edc754f142ac-kube-api-access-s448x\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379500 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-combined-ca-bundle\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379592 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-ssh-key\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379662 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-0\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379749 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-0\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.379792 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-inventory\") pod \"fc8364bf-1ce5-4681-99a5-edc754f142ac\" (UID: \"fc8364bf-1ce5-4681-99a5-edc754f142ac\") " Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.389025 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.403173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8364bf-1ce5-4681-99a5-edc754f142ac-kube-api-access-s448x" (OuterVolumeSpecName: "kube-api-access-s448x") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "kube-api-access-s448x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.406991 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.412094 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.413800 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.413913 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.415398 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-inventory" (OuterVolumeSpecName: "inventory") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.422686 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.439906 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "fc8364bf-1ce5-4681-99a5-edc754f142ac" (UID: "fc8364bf-1ce5-4681-99a5-edc754f142ac"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.481998 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482043 4775 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482056 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482070 4775 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482082 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s448x\" (UniqueName: \"kubernetes.io/projected/fc8364bf-1ce5-4681-99a5-edc754f142ac-kube-api-access-s448x\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482096 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482108 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482120 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.482132 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc8364bf-1ce5-4681-99a5-edc754f142ac-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.766012 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.779653 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-ldl57" event={"ID":"fc8364bf-1ce5-4681-99a5-edc754f142ac","Type":"ContainerDied","Data":"f1b8972c5bbf2d1c670c32acf07ca62a0560784f02334c72d4892c0494126f75"} Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.779752 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b8972c5bbf2d1c670c32acf07ca62a0560784f02334c72d4892c0494126f75" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.863367 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps"] Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.863925 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="registry-server" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.863944 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="registry-server" Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.863968 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="registry-server" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.863978 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="registry-server" Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.863996 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="extract-content" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864006 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="extract-content" Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.864019 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc8364bf-1ce5-4681-99a5-edc754f142ac" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864028 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc8364bf-1ce5-4681-99a5-edc754f142ac" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.864038 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="extract-content" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864045 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="extract-content" Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.864061 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="extract-utilities" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864068 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="extract-utilities" Nov 26 07:05:13 crc kubenswrapper[4775]: E1126 07:05:13.864094 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="extract-utilities" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864104 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="extract-utilities" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864332 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="514b3d70-4c1d-47a4-b700-5c4381ee371d" containerName="registry-server" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864346 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc8364bf-1ce5-4681-99a5-edc754f142ac" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.864361 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7993caa1-b677-44ba-a755-26f4c130fab2" containerName="registry-server" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.865214 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.873235 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.873477 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.873705 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.873864 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-nlzkt" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.874007 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.895975 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps"] Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.991867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.992178 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.992335 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.992472 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js4tc\" (UniqueName: \"kubernetes.io/projected/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-kube-api-access-js4tc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.992611 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.992760 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:13 crc kubenswrapper[4775]: I1126 07:05:13.992859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.094569 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.095171 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.095496 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.095785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.096121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.096434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js4tc\" (UniqueName: \"kubernetes.io/projected/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-kube-api-access-js4tc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.096835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.098157 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.100200 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.100785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.101137 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.102992 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.107317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.125406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js4tc\" (UniqueName: \"kubernetes.io/projected/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-kube-api-access-js4tc\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ssps\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.191466 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.676472 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps"] Nov 26 07:05:14 crc kubenswrapper[4775]: I1126 07:05:14.785486 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" event={"ID":"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf","Type":"ContainerStarted","Data":"a55c8fdce3aea66a3fb83992f768afac46ad8c72c7023d7767f040a54f50dac5"} Nov 26 07:05:15 crc kubenswrapper[4775]: I1126 07:05:15.793912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" event={"ID":"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf","Type":"ContainerStarted","Data":"42818833f04c3b35c7093de8c4888ed34bbfeda25bebbb85fa11b29a938dca3f"} Nov 26 07:05:15 crc kubenswrapper[4775]: I1126 07:05:15.814790 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" podStartSLOduration=2.36273516 podStartE2EDuration="2.814773921s" podCreationTimestamp="2025-11-26 07:05:13 +0000 UTC" firstStartedPulling="2025-11-26 07:05:14.682853086 +0000 UTC m=+2398.044157038" lastFinishedPulling="2025-11-26 07:05:15.134891807 +0000 UTC m=+2398.496195799" observedRunningTime="2025-11-26 07:05:15.812571883 +0000 UTC m=+2399.173875835" watchObservedRunningTime="2025-11-26 07:05:15.814773921 +0000 UTC m=+2399.176077873" Nov 26 07:05:23 crc kubenswrapper[4775]: I1126 07:05:23.328687 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:05:23 crc kubenswrapper[4775]: E1126 07:05:23.329581 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:05:38 crc kubenswrapper[4775]: I1126 07:05:38.329903 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:05:38 crc kubenswrapper[4775]: E1126 07:05:38.330598 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:05:51 crc kubenswrapper[4775]: I1126 07:05:51.328098 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:05:51 crc kubenswrapper[4775]: E1126 07:05:51.330783 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:06:04 crc kubenswrapper[4775]: I1126 07:06:04.328387 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:06:04 crc kubenswrapper[4775]: E1126 07:06:04.329252 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:06:17 crc kubenswrapper[4775]: I1126 07:06:17.334323 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:06:17 crc kubenswrapper[4775]: E1126 07:06:17.335156 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:06:31 crc kubenswrapper[4775]: I1126 07:06:31.327586 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:06:32 crc kubenswrapper[4775]: I1126 07:06:32.487637 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"a37665ebfb8085298b35742ecbe705bad042a529c1a8da9d519567710c7c8f62"} Nov 26 07:07:45 crc kubenswrapper[4775]: I1126 07:07:45.152472 4775 generic.go:334] "Generic (PLEG): container finished" podID="1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" containerID="42818833f04c3b35c7093de8c4888ed34bbfeda25bebbb85fa11b29a938dca3f" exitCode=0 Nov 26 07:07:45 crc kubenswrapper[4775]: I1126 07:07:45.152556 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" event={"ID":"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf","Type":"ContainerDied","Data":"42818833f04c3b35c7093de8c4888ed34bbfeda25bebbb85fa11b29a938dca3f"} Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.602448 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.739981 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js4tc\" (UniqueName: \"kubernetes.io/projected/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-kube-api-access-js4tc\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.740043 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-telemetry-combined-ca-bundle\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.740182 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-1\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.740234 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ssh-key\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.740261 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-0\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.740298 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-inventory\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.740351 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-2\") pod \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\" (UID: \"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf\") " Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.746835 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-kube-api-access-js4tc" (OuterVolumeSpecName: "kube-api-access-js4tc") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "kube-api-access-js4tc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.747637 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.773494 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.776119 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.777659 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.780537 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-inventory" (OuterVolumeSpecName: "inventory") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.782018 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" (UID: "1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843004 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js4tc\" (UniqueName: \"kubernetes.io/projected/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-kube-api-access-js4tc\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843045 4775 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843060 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843074 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843087 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843100 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-inventory\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:46 crc kubenswrapper[4775]: I1126 07:07:46.843111 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 26 07:07:47 crc kubenswrapper[4775]: I1126 07:07:47.171077 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" event={"ID":"1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf","Type":"ContainerDied","Data":"a55c8fdce3aea66a3fb83992f768afac46ad8c72c7023d7767f040a54f50dac5"} Nov 26 07:07:47 crc kubenswrapper[4775]: I1126 07:07:47.171407 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a55c8fdce3aea66a3fb83992f768afac46ad8c72c7023d7767f040a54f50dac5" Nov 26 07:07:47 crc kubenswrapper[4775]: I1126 07:07:47.171128 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ssps" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.551987 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 26 07:08:47 crc kubenswrapper[4775]: E1126 07:08:47.553099 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.553120 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.553419 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.554338 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.556488 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5p94c" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.557123 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.557615 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.558532 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.563443 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.634859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.634915 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msvkm\" (UniqueName: \"kubernetes.io/projected/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-kube-api-access-msvkm\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.634958 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.635100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.635142 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-config-data\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.635254 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.635418 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.635524 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.635555 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737221 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msvkm\" (UniqueName: \"kubernetes.io/projected/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-kube-api-access-msvkm\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737305 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737407 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737437 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-config-data\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737584 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737606 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.737635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.738173 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.738702 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-config-data\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.738774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.739003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.739681 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.745277 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.745564 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.753512 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.758498 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msvkm\" (UniqueName: \"kubernetes.io/projected/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-kube-api-access-msvkm\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.785029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " pod="openstack/tempest-tests-tempest" Nov 26 07:08:47 crc kubenswrapper[4775]: I1126 07:08:47.874808 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 07:08:48 crc kubenswrapper[4775]: I1126 07:08:48.329888 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 26 07:08:48 crc kubenswrapper[4775]: I1126 07:08:48.334295 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 07:08:48 crc kubenswrapper[4775]: I1126 07:08:48.817911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5","Type":"ContainerStarted","Data":"400ef64c5ebb828ab92a46428f5aff3daccc116ad55b31d079d9692f2972ce63"} Nov 26 07:08:51 crc kubenswrapper[4775]: I1126 07:08:51.420232 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:08:51 crc kubenswrapper[4775]: I1126 07:08:51.420687 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.309198 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tl76k"] Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.312096 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.326189 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tl76k"] Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.455515 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp9m5\" (UniqueName: \"kubernetes.io/projected/7a0a4545-8d51-496e-9d29-ef2aa622d467-kube-api-access-hp9m5\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.455646 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-catalog-content\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.455788 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-utilities\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.557830 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-utilities\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.557941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp9m5\" (UniqueName: \"kubernetes.io/projected/7a0a4545-8d51-496e-9d29-ef2aa622d467-kube-api-access-hp9m5\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.557991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-catalog-content\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.558375 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-utilities\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.558422 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-catalog-content\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.586377 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp9m5\" (UniqueName: \"kubernetes.io/projected/7a0a4545-8d51-496e-9d29-ef2aa622d467-kube-api-access-hp9m5\") pod \"certified-operators-tl76k\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:19 crc kubenswrapper[4775]: I1126 07:09:19.643673 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:21 crc kubenswrapper[4775]: I1126 07:09:21.419510 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:09:21 crc kubenswrapper[4775]: I1126 07:09:21.419833 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:09:22 crc kubenswrapper[4775]: E1126 07:09:22.052696 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Nov 26 07:09:22 crc kubenswrapper[4775]: E1126 07:09:22.053131 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msvkm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 07:09:22 crc kubenswrapper[4775]: E1126 07:09:22.055124 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" Nov 26 07:09:22 crc kubenswrapper[4775]: E1126 07:09:22.126753 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" Nov 26 07:09:22 crc kubenswrapper[4775]: I1126 07:09:22.312549 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tl76k"] Nov 26 07:09:23 crc kubenswrapper[4775]: I1126 07:09:23.134989 4775 generic.go:334] "Generic (PLEG): container finished" podID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerID="5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a" exitCode=0 Nov 26 07:09:23 crc kubenswrapper[4775]: I1126 07:09:23.135086 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerDied","Data":"5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a"} Nov 26 07:09:23 crc kubenswrapper[4775]: I1126 07:09:23.135292 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerStarted","Data":"3d12e35df1f7e82d648b5ca5d2e35157e2bed978fe4bb94aec2eb986b966e5b2"} Nov 26 07:09:24 crc kubenswrapper[4775]: I1126 07:09:24.145910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerStarted","Data":"4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96"} Nov 26 07:09:25 crc kubenswrapper[4775]: I1126 07:09:25.155997 4775 generic.go:334] "Generic (PLEG): container finished" podID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerID="4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96" exitCode=0 Nov 26 07:09:25 crc kubenswrapper[4775]: I1126 07:09:25.156075 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerDied","Data":"4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96"} Nov 26 07:09:26 crc kubenswrapper[4775]: I1126 07:09:26.166961 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerStarted","Data":"80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63"} Nov 26 07:09:26 crc kubenswrapper[4775]: I1126 07:09:26.189532 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tl76k" podStartSLOduration=4.789378039 podStartE2EDuration="7.189511518s" podCreationTimestamp="2025-11-26 07:09:19 +0000 UTC" firstStartedPulling="2025-11-26 07:09:23.138105571 +0000 UTC m=+2646.499409523" lastFinishedPulling="2025-11-26 07:09:25.53823905 +0000 UTC m=+2648.899543002" observedRunningTime="2025-11-26 07:09:26.182844062 +0000 UTC m=+2649.544148024" watchObservedRunningTime="2025-11-26 07:09:26.189511518 +0000 UTC m=+2649.550815470" Nov 26 07:09:29 crc kubenswrapper[4775]: I1126 07:09:29.644967 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:29 crc kubenswrapper[4775]: I1126 07:09:29.645290 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:29 crc kubenswrapper[4775]: I1126 07:09:29.687432 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:30 crc kubenswrapper[4775]: I1126 07:09:30.278900 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:30 crc kubenswrapper[4775]: I1126 07:09:30.326577 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tl76k"] Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.242656 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tl76k" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="registry-server" containerID="cri-o://80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63" gracePeriod=2 Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.668901 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.812795 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-catalog-content\") pod \"7a0a4545-8d51-496e-9d29-ef2aa622d467\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.813484 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp9m5\" (UniqueName: \"kubernetes.io/projected/7a0a4545-8d51-496e-9d29-ef2aa622d467-kube-api-access-hp9m5\") pod \"7a0a4545-8d51-496e-9d29-ef2aa622d467\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.813521 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-utilities\") pod \"7a0a4545-8d51-496e-9d29-ef2aa622d467\" (UID: \"7a0a4545-8d51-496e-9d29-ef2aa622d467\") " Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.814305 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-utilities" (OuterVolumeSpecName: "utilities") pod "7a0a4545-8d51-496e-9d29-ef2aa622d467" (UID: "7a0a4545-8d51-496e-9d29-ef2aa622d467"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.826550 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a0a4545-8d51-496e-9d29-ef2aa622d467-kube-api-access-hp9m5" (OuterVolumeSpecName: "kube-api-access-hp9m5") pod "7a0a4545-8d51-496e-9d29-ef2aa622d467" (UID: "7a0a4545-8d51-496e-9d29-ef2aa622d467"). InnerVolumeSpecName "kube-api-access-hp9m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.867766 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a0a4545-8d51-496e-9d29-ef2aa622d467" (UID: "7a0a4545-8d51-496e-9d29-ef2aa622d467"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.916137 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp9m5\" (UniqueName: \"kubernetes.io/projected/7a0a4545-8d51-496e-9d29-ef2aa622d467-kube-api-access-hp9m5\") on node \"crc\" DevicePath \"\"" Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.916181 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:09:32 crc kubenswrapper[4775]: I1126 07:09:32.916193 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a0a4545-8d51-496e-9d29-ef2aa622d467-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.252233 4775 generic.go:334] "Generic (PLEG): container finished" podID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerID="80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63" exitCode=0 Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.252276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerDied","Data":"80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63"} Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.252290 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tl76k" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.252302 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tl76k" event={"ID":"7a0a4545-8d51-496e-9d29-ef2aa622d467","Type":"ContainerDied","Data":"3d12e35df1f7e82d648b5ca5d2e35157e2bed978fe4bb94aec2eb986b966e5b2"} Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.252318 4775 scope.go:117] "RemoveContainer" containerID="80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.286796 4775 scope.go:117] "RemoveContainer" containerID="4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.293489 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tl76k"] Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.307198 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tl76k"] Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.307555 4775 scope.go:117] "RemoveContainer" containerID="5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.336797 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" path="/var/lib/kubelet/pods/7a0a4545-8d51-496e-9d29-ef2aa622d467/volumes" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.368174 4775 scope.go:117] "RemoveContainer" containerID="80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63" Nov 26 07:09:33 crc kubenswrapper[4775]: E1126 07:09:33.370136 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63\": container with ID starting with 80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63 not found: ID does not exist" containerID="80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.370186 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63"} err="failed to get container status \"80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63\": rpc error: code = NotFound desc = could not find container \"80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63\": container with ID starting with 80756f6d3ecdafef4aadbf2ef3751b651ffd8830372c4368692a6dc87f000a63 not found: ID does not exist" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.370221 4775 scope.go:117] "RemoveContainer" containerID="4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96" Nov 26 07:09:33 crc kubenswrapper[4775]: E1126 07:09:33.370895 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96\": container with ID starting with 4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96 not found: ID does not exist" containerID="4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.370939 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96"} err="failed to get container status \"4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96\": rpc error: code = NotFound desc = could not find container \"4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96\": container with ID starting with 4cda6cc4099b56a375c3d22410c875937e3d71138f44af145aef372b77cece96 not found: ID does not exist" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.370966 4775 scope.go:117] "RemoveContainer" containerID="5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a" Nov 26 07:09:33 crc kubenswrapper[4775]: E1126 07:09:33.372197 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a\": container with ID starting with 5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a not found: ID does not exist" containerID="5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a" Nov 26 07:09:33 crc kubenswrapper[4775]: I1126 07:09:33.372240 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a"} err="failed to get container status \"5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a\": rpc error: code = NotFound desc = could not find container \"5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a\": container with ID starting with 5e19bb82adf63ca38b49d91de2835b23d6d6aa6f4e38e40f410679b649cff60a not found: ID does not exist" Nov 26 07:09:37 crc kubenswrapper[4775]: I1126 07:09:37.816741 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 26 07:09:39 crc kubenswrapper[4775]: I1126 07:09:39.317892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5","Type":"ContainerStarted","Data":"3e57123e18cda991cc838d510f81306094d91df222c84e96f42e9d4be423739e"} Nov 26 07:09:39 crc kubenswrapper[4775]: I1126 07:09:39.346421 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.865907309 podStartE2EDuration="53.346403932s" podCreationTimestamp="2025-11-26 07:08:46 +0000 UTC" firstStartedPulling="2025-11-26 07:08:48.33408065 +0000 UTC m=+2611.695384602" lastFinishedPulling="2025-11-26 07:09:37.814577273 +0000 UTC m=+2661.175881225" observedRunningTime="2025-11-26 07:09:39.33993141 +0000 UTC m=+2662.701235372" watchObservedRunningTime="2025-11-26 07:09:39.346403932 +0000 UTC m=+2662.707707894" Nov 26 07:09:51 crc kubenswrapper[4775]: I1126 07:09:51.419896 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:09:51 crc kubenswrapper[4775]: I1126 07:09:51.420626 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:09:51 crc kubenswrapper[4775]: I1126 07:09:51.420706 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:09:51 crc kubenswrapper[4775]: I1126 07:09:51.421880 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a37665ebfb8085298b35742ecbe705bad042a529c1a8da9d519567710c7c8f62"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:09:51 crc kubenswrapper[4775]: I1126 07:09:51.421987 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://a37665ebfb8085298b35742ecbe705bad042a529c1a8da9d519567710c7c8f62" gracePeriod=600 Nov 26 07:09:52 crc kubenswrapper[4775]: I1126 07:09:52.462077 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="a37665ebfb8085298b35742ecbe705bad042a529c1a8da9d519567710c7c8f62" exitCode=0 Nov 26 07:09:52 crc kubenswrapper[4775]: I1126 07:09:52.462226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"a37665ebfb8085298b35742ecbe705bad042a529c1a8da9d519567710c7c8f62"} Nov 26 07:09:52 crc kubenswrapper[4775]: I1126 07:09:52.462781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5"} Nov 26 07:09:52 crc kubenswrapper[4775]: I1126 07:09:52.462819 4775 scope.go:117] "RemoveContainer" containerID="34cd919dd0c692c433efc9f9a1c2829c3e83ae00b0fce99c0524bd2c43ace689" Nov 26 07:11:51 crc kubenswrapper[4775]: I1126 07:11:51.419611 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:11:51 crc kubenswrapper[4775]: I1126 07:11:51.420137 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:12:21 crc kubenswrapper[4775]: I1126 07:12:21.419485 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:12:21 crc kubenswrapper[4775]: I1126 07:12:21.419970 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.713874 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zhr46"] Nov 26 07:12:41 crc kubenswrapper[4775]: E1126 07:12:41.714878 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="registry-server" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.714897 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="registry-server" Nov 26 07:12:41 crc kubenswrapper[4775]: E1126 07:12:41.714916 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="extract-content" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.714923 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="extract-content" Nov 26 07:12:41 crc kubenswrapper[4775]: E1126 07:12:41.714954 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="extract-utilities" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.714962 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="extract-utilities" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.715161 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a0a4545-8d51-496e-9d29-ef2aa622d467" containerName="registry-server" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.716619 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.727542 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhr46"] Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.846483 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-utilities\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.846552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdswh\" (UniqueName: \"kubernetes.io/projected/ec894291-5646-4f5e-9a9e-e509830af6d9-kube-api-access-qdswh\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.846876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-catalog-content\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.948433 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-utilities\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.948510 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdswh\" (UniqueName: \"kubernetes.io/projected/ec894291-5646-4f5e-9a9e-e509830af6d9-kube-api-access-qdswh\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.948625 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-catalog-content\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.949081 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-catalog-content\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.949176 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-utilities\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:41 crc kubenswrapper[4775]: I1126 07:12:41.972888 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdswh\" (UniqueName: \"kubernetes.io/projected/ec894291-5646-4f5e-9a9e-e509830af6d9-kube-api-access-qdswh\") pod \"redhat-marketplace-zhr46\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:42 crc kubenswrapper[4775]: I1126 07:12:42.045143 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:42 crc kubenswrapper[4775]: I1126 07:12:42.516827 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhr46"] Nov 26 07:12:42 crc kubenswrapper[4775]: I1126 07:12:42.984910 4775 generic.go:334] "Generic (PLEG): container finished" podID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerID="978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119" exitCode=0 Nov 26 07:12:42 crc kubenswrapper[4775]: I1126 07:12:42.985972 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerDied","Data":"978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119"} Nov 26 07:12:42 crc kubenswrapper[4775]: I1126 07:12:42.986078 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerStarted","Data":"01a3277dc0f603f6af095b543d9e334b7868aa234f9fc3b93e72bc3300eb5691"} Nov 26 07:12:43 crc kubenswrapper[4775]: I1126 07:12:43.994881 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerStarted","Data":"7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5"} Nov 26 07:12:45 crc kubenswrapper[4775]: I1126 07:12:45.006554 4775 generic.go:334] "Generic (PLEG): container finished" podID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerID="7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5" exitCode=0 Nov 26 07:12:45 crc kubenswrapper[4775]: I1126 07:12:45.006672 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerDied","Data":"7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5"} Nov 26 07:12:46 crc kubenswrapper[4775]: I1126 07:12:46.017781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerStarted","Data":"5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f"} Nov 26 07:12:46 crc kubenswrapper[4775]: I1126 07:12:46.036546 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zhr46" podStartSLOduration=2.614502884 podStartE2EDuration="5.036522753s" podCreationTimestamp="2025-11-26 07:12:41 +0000 UTC" firstStartedPulling="2025-11-26 07:12:42.986991787 +0000 UTC m=+2846.348295739" lastFinishedPulling="2025-11-26 07:12:45.409011656 +0000 UTC m=+2848.770315608" observedRunningTime="2025-11-26 07:12:46.033493553 +0000 UTC m=+2849.394797545" watchObservedRunningTime="2025-11-26 07:12:46.036522753 +0000 UTC m=+2849.397826725" Nov 26 07:12:51 crc kubenswrapper[4775]: I1126 07:12:51.420218 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:12:51 crc kubenswrapper[4775]: I1126 07:12:51.420775 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:12:51 crc kubenswrapper[4775]: I1126 07:12:51.420826 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:12:51 crc kubenswrapper[4775]: I1126 07:12:51.421575 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:12:51 crc kubenswrapper[4775]: I1126 07:12:51.421631 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" gracePeriod=600 Nov 26 07:12:51 crc kubenswrapper[4775]: E1126 07:12:51.552519 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.048312 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.048372 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.090234 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" exitCode=0 Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.090272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5"} Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.090305 4775 scope.go:117] "RemoveContainer" containerID="a37665ebfb8085298b35742ecbe705bad042a529c1a8da9d519567710c7c8f62" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.090977 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:12:52 crc kubenswrapper[4775]: E1126 07:12:52.091441 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.096892 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.155464 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.698736 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7v2ns"] Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.701067 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.719039 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7v2ns"] Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.764846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-catalog-content\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.764983 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9slz\" (UniqueName: \"kubernetes.io/projected/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-kube-api-access-k9slz\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.765084 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-utilities\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.867466 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-catalog-content\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.867587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9slz\" (UniqueName: \"kubernetes.io/projected/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-kube-api-access-k9slz\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.867662 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-utilities\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.867975 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-catalog-content\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.867987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-utilities\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:52 crc kubenswrapper[4775]: I1126 07:12:52.888027 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9slz\" (UniqueName: \"kubernetes.io/projected/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-kube-api-access-k9slz\") pod \"community-operators-7v2ns\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:53 crc kubenswrapper[4775]: I1126 07:12:53.034174 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:12:53 crc kubenswrapper[4775]: I1126 07:12:53.576804 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7v2ns"] Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.090090 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhr46"] Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.130205 4775 generic.go:334] "Generic (PLEG): container finished" podID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerID="f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f" exitCode=0 Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.130265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerDied","Data":"f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f"} Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.130318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerStarted","Data":"1b7371248f11ffc6d7cbe337b78f212701067834ad812517dedfa6fd56f3dd88"} Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.130545 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zhr46" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="registry-server" containerID="cri-o://5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f" gracePeriod=2 Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.664560 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.803039 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-catalog-content\") pod \"ec894291-5646-4f5e-9a9e-e509830af6d9\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.803143 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdswh\" (UniqueName: \"kubernetes.io/projected/ec894291-5646-4f5e-9a9e-e509830af6d9-kube-api-access-qdswh\") pod \"ec894291-5646-4f5e-9a9e-e509830af6d9\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.803346 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-utilities\") pod \"ec894291-5646-4f5e-9a9e-e509830af6d9\" (UID: \"ec894291-5646-4f5e-9a9e-e509830af6d9\") " Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.804086 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-utilities" (OuterVolumeSpecName: "utilities") pod "ec894291-5646-4f5e-9a9e-e509830af6d9" (UID: "ec894291-5646-4f5e-9a9e-e509830af6d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.812088 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec894291-5646-4f5e-9a9e-e509830af6d9-kube-api-access-qdswh" (OuterVolumeSpecName: "kube-api-access-qdswh") pod "ec894291-5646-4f5e-9a9e-e509830af6d9" (UID: "ec894291-5646-4f5e-9a9e-e509830af6d9"). InnerVolumeSpecName "kube-api-access-qdswh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.822487 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec894291-5646-4f5e-9a9e-e509830af6d9" (UID: "ec894291-5646-4f5e-9a9e-e509830af6d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.905872 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.906240 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdswh\" (UniqueName: \"kubernetes.io/projected/ec894291-5646-4f5e-9a9e-e509830af6d9-kube-api-access-qdswh\") on node \"crc\" DevicePath \"\"" Nov 26 07:12:54 crc kubenswrapper[4775]: I1126 07:12:54.906253 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec894291-5646-4f5e-9a9e-e509830af6d9-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.142327 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerStarted","Data":"2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034"} Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.144367 4775 generic.go:334] "Generic (PLEG): container finished" podID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerID="5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f" exitCode=0 Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.144422 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zhr46" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.144418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerDied","Data":"5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f"} Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.144572 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zhr46" event={"ID":"ec894291-5646-4f5e-9a9e-e509830af6d9","Type":"ContainerDied","Data":"01a3277dc0f603f6af095b543d9e334b7868aa234f9fc3b93e72bc3300eb5691"} Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.144610 4775 scope.go:117] "RemoveContainer" containerID="5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.188331 4775 scope.go:117] "RemoveContainer" containerID="7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.189836 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhr46"] Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.207391 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zhr46"] Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.209278 4775 scope.go:117] "RemoveContainer" containerID="978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.258711 4775 scope.go:117] "RemoveContainer" containerID="5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f" Nov 26 07:12:55 crc kubenswrapper[4775]: E1126 07:12:55.259304 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f\": container with ID starting with 5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f not found: ID does not exist" containerID="5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.259355 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f"} err="failed to get container status \"5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f\": rpc error: code = NotFound desc = could not find container \"5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f\": container with ID starting with 5704f28627bfdc678e4c5ebe100fa84c1eb3618e72d273ebf4fd63ee3d00ac8f not found: ID does not exist" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.259386 4775 scope.go:117] "RemoveContainer" containerID="7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5" Nov 26 07:12:55 crc kubenswrapper[4775]: E1126 07:12:55.259648 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5\": container with ID starting with 7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5 not found: ID does not exist" containerID="7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.259675 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5"} err="failed to get container status \"7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5\": rpc error: code = NotFound desc = could not find container \"7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5\": container with ID starting with 7058ac1e2456d49452d23b5f90970d5c4e95040b94542d67e4ffdf4fa8d103f5 not found: ID does not exist" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.259690 4775 scope.go:117] "RemoveContainer" containerID="978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119" Nov 26 07:12:55 crc kubenswrapper[4775]: E1126 07:12:55.259996 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119\": container with ID starting with 978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119 not found: ID does not exist" containerID="978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.260021 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119"} err="failed to get container status \"978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119\": rpc error: code = NotFound desc = could not find container \"978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119\": container with ID starting with 978a882ed82bde79608c6030818a435520a078566f1f2b6a820c3938cc566119 not found: ID does not exist" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.338991 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" path="/var/lib/kubelet/pods/ec894291-5646-4f5e-9a9e-e509830af6d9/volumes" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.500534 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c9vsk"] Nov 26 07:12:55 crc kubenswrapper[4775]: E1126 07:12:55.501190 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="extract-utilities" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.501228 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="extract-utilities" Nov 26 07:12:55 crc kubenswrapper[4775]: E1126 07:12:55.501310 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="registry-server" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.501332 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="registry-server" Nov 26 07:12:55 crc kubenswrapper[4775]: E1126 07:12:55.501380 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="extract-content" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.501394 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="extract-content" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.501755 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec894291-5646-4f5e-9a9e-e509830af6d9" containerName="registry-server" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.504460 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.521603 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-utilities\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.521888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-catalog-content\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.521936 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx7l2\" (UniqueName: \"kubernetes.io/projected/3894718b-9c8e-4256-8f71-9bfe377bac10-kube-api-access-sx7l2\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.526488 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c9vsk"] Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.623841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-utilities\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.623963 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-catalog-content\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.623991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx7l2\" (UniqueName: \"kubernetes.io/projected/3894718b-9c8e-4256-8f71-9bfe377bac10-kube-api-access-sx7l2\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.624337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-utilities\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.624374 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-catalog-content\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.641234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx7l2\" (UniqueName: \"kubernetes.io/projected/3894718b-9c8e-4256-8f71-9bfe377bac10-kube-api-access-sx7l2\") pod \"redhat-operators-c9vsk\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:55 crc kubenswrapper[4775]: I1126 07:12:55.837775 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:12:56 crc kubenswrapper[4775]: I1126 07:12:56.159857 4775 generic.go:334] "Generic (PLEG): container finished" podID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerID="2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034" exitCode=0 Nov 26 07:12:56 crc kubenswrapper[4775]: I1126 07:12:56.159977 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerDied","Data":"2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034"} Nov 26 07:12:56 crc kubenswrapper[4775]: I1126 07:12:56.319186 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c9vsk"] Nov 26 07:12:56 crc kubenswrapper[4775]: W1126 07:12:56.322266 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3894718b_9c8e_4256_8f71_9bfe377bac10.slice/crio-93b174f0293e82c407f67c773437c9a34c527eaf960677aca48663c04488707d WatchSource:0}: Error finding container 93b174f0293e82c407f67c773437c9a34c527eaf960677aca48663c04488707d: Status 404 returned error can't find the container with id 93b174f0293e82c407f67c773437c9a34c527eaf960677aca48663c04488707d Nov 26 07:12:57 crc kubenswrapper[4775]: I1126 07:12:57.178080 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerStarted","Data":"93b174f0293e82c407f67c773437c9a34c527eaf960677aca48663c04488707d"} Nov 26 07:12:58 crc kubenswrapper[4775]: I1126 07:12:58.191855 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerStarted","Data":"45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9"} Nov 26 07:12:58 crc kubenswrapper[4775]: I1126 07:12:58.194209 4775 generic.go:334] "Generic (PLEG): container finished" podID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerID="cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889" exitCode=0 Nov 26 07:12:58 crc kubenswrapper[4775]: I1126 07:12:58.194251 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerDied","Data":"cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889"} Nov 26 07:12:58 crc kubenswrapper[4775]: I1126 07:12:58.238146 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7v2ns" podStartSLOduration=2.521500606 podStartE2EDuration="6.238129911s" podCreationTimestamp="2025-11-26 07:12:52 +0000 UTC" firstStartedPulling="2025-11-26 07:12:54.132953331 +0000 UTC m=+2857.494257283" lastFinishedPulling="2025-11-26 07:12:57.849582636 +0000 UTC m=+2861.210886588" observedRunningTime="2025-11-26 07:12:58.222237608 +0000 UTC m=+2861.583541560" watchObservedRunningTime="2025-11-26 07:12:58.238129911 +0000 UTC m=+2861.599433853" Nov 26 07:12:59 crc kubenswrapper[4775]: I1126 07:12:59.206011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerStarted","Data":"11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035"} Nov 26 07:13:02 crc kubenswrapper[4775]: I1126 07:13:02.240787 4775 generic.go:334] "Generic (PLEG): container finished" podID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerID="11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035" exitCode=0 Nov 26 07:13:02 crc kubenswrapper[4775]: I1126 07:13:02.240920 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerDied","Data":"11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035"} Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.035422 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.035768 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.079339 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.253897 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerStarted","Data":"38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2"} Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.273735 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c9vsk" podStartSLOduration=3.812894054 podStartE2EDuration="8.273695263s" podCreationTimestamp="2025-11-26 07:12:55 +0000 UTC" firstStartedPulling="2025-11-26 07:12:58.196961335 +0000 UTC m=+2861.558265287" lastFinishedPulling="2025-11-26 07:13:02.657762524 +0000 UTC m=+2866.019066496" observedRunningTime="2025-11-26 07:13:03.271254048 +0000 UTC m=+2866.632558020" watchObservedRunningTime="2025-11-26 07:13:03.273695263 +0000 UTC m=+2866.634999215" Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.312240 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:13:03 crc kubenswrapper[4775]: I1126 07:13:03.327689 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:13:03 crc kubenswrapper[4775]: E1126 07:13:03.328005 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.288295 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7v2ns"] Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.288514 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7v2ns" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="registry-server" containerID="cri-o://45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9" gracePeriod=2 Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.838043 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.838318 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.848231 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.955042 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9slz\" (UniqueName: \"kubernetes.io/projected/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-kube-api-access-k9slz\") pod \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.955102 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-utilities\") pod \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.955253 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-catalog-content\") pod \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\" (UID: \"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99\") " Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.955898 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-utilities" (OuterVolumeSpecName: "utilities") pod "48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" (UID: "48a67f4c-f3c9-4ef8-8224-5e1ab3944f99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:13:05 crc kubenswrapper[4775]: I1126 07:13:05.963942 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-kube-api-access-k9slz" (OuterVolumeSpecName: "kube-api-access-k9slz") pod "48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" (UID: "48a67f4c-f3c9-4ef8-8224-5e1ab3944f99"). InnerVolumeSpecName "kube-api-access-k9slz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.003313 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" (UID: "48a67f4c-f3c9-4ef8-8224-5e1ab3944f99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.058796 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.058837 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9slz\" (UniqueName: \"kubernetes.io/projected/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-kube-api-access-k9slz\") on node \"crc\" DevicePath \"\"" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.058852 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.283370 4775 generic.go:334] "Generic (PLEG): container finished" podID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerID="45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9" exitCode=0 Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.283444 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerDied","Data":"45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9"} Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.283497 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7v2ns" event={"ID":"48a67f4c-f3c9-4ef8-8224-5e1ab3944f99","Type":"ContainerDied","Data":"1b7371248f11ffc6d7cbe337b78f212701067834ad812517dedfa6fd56f3dd88"} Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.283519 4775 scope.go:117] "RemoveContainer" containerID="45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.283738 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7v2ns" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.314981 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7v2ns"] Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.318026 4775 scope.go:117] "RemoveContainer" containerID="2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.322980 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7v2ns"] Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.339371 4775 scope.go:117] "RemoveContainer" containerID="f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.389955 4775 scope.go:117] "RemoveContainer" containerID="45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9" Nov 26 07:13:06 crc kubenswrapper[4775]: E1126 07:13:06.390350 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9\": container with ID starting with 45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9 not found: ID does not exist" containerID="45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.390402 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9"} err="failed to get container status \"45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9\": rpc error: code = NotFound desc = could not find container \"45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9\": container with ID starting with 45f4a538fcfd8182d63cdedb2b46d20df626ed37264153b889f982a0240c69b9 not found: ID does not exist" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.390423 4775 scope.go:117] "RemoveContainer" containerID="2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034" Nov 26 07:13:06 crc kubenswrapper[4775]: E1126 07:13:06.390701 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034\": container with ID starting with 2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034 not found: ID does not exist" containerID="2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.390746 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034"} err="failed to get container status \"2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034\": rpc error: code = NotFound desc = could not find container \"2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034\": container with ID starting with 2a6ce35a255557c8d766f46b12d4d73babb508ee3288f8d66d4c3024a5fa7034 not found: ID does not exist" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.390760 4775 scope.go:117] "RemoveContainer" containerID="f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f" Nov 26 07:13:06 crc kubenswrapper[4775]: E1126 07:13:06.391134 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f\": container with ID starting with f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f not found: ID does not exist" containerID="f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.391154 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f"} err="failed to get container status \"f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f\": rpc error: code = NotFound desc = could not find container \"f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f\": container with ID starting with f7d9068fa0d97a1e347f3ec3298ebed0d890309aa5f7597852555b1420f3309f not found: ID does not exist" Nov 26 07:13:06 crc kubenswrapper[4775]: I1126 07:13:06.899532 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c9vsk" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="registry-server" probeResult="failure" output=< Nov 26 07:13:06 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Nov 26 07:13:06 crc kubenswrapper[4775]: > Nov 26 07:13:07 crc kubenswrapper[4775]: I1126 07:13:07.354245 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" path="/var/lib/kubelet/pods/48a67f4c-f3c9-4ef8-8224-5e1ab3944f99/volumes" Nov 26 07:13:14 crc kubenswrapper[4775]: I1126 07:13:14.328486 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:13:14 crc kubenswrapper[4775]: E1126 07:13:14.329513 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:13:15 crc kubenswrapper[4775]: I1126 07:13:15.884843 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:13:15 crc kubenswrapper[4775]: I1126 07:13:15.936400 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.292628 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c9vsk"] Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.293573 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c9vsk" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="registry-server" containerID="cri-o://38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2" gracePeriod=2 Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.785109 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.906315 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-utilities\") pod \"3894718b-9c8e-4256-8f71-9bfe377bac10\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.906543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx7l2\" (UniqueName: \"kubernetes.io/projected/3894718b-9c8e-4256-8f71-9bfe377bac10-kube-api-access-sx7l2\") pod \"3894718b-9c8e-4256-8f71-9bfe377bac10\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.906586 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-catalog-content\") pod \"3894718b-9c8e-4256-8f71-9bfe377bac10\" (UID: \"3894718b-9c8e-4256-8f71-9bfe377bac10\") " Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.907509 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-utilities" (OuterVolumeSpecName: "utilities") pod "3894718b-9c8e-4256-8f71-9bfe377bac10" (UID: "3894718b-9c8e-4256-8f71-9bfe377bac10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:13:19 crc kubenswrapper[4775]: I1126 07:13:19.916562 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3894718b-9c8e-4256-8f71-9bfe377bac10-kube-api-access-sx7l2" (OuterVolumeSpecName: "kube-api-access-sx7l2") pod "3894718b-9c8e-4256-8f71-9bfe377bac10" (UID: "3894718b-9c8e-4256-8f71-9bfe377bac10"). InnerVolumeSpecName "kube-api-access-sx7l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.002010 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3894718b-9c8e-4256-8f71-9bfe377bac10" (UID: "3894718b-9c8e-4256-8f71-9bfe377bac10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.008551 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.008592 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx7l2\" (UniqueName: \"kubernetes.io/projected/3894718b-9c8e-4256-8f71-9bfe377bac10-kube-api-access-sx7l2\") on node \"crc\" DevicePath \"\"" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.008606 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3894718b-9c8e-4256-8f71-9bfe377bac10-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.426004 4775 generic.go:334] "Generic (PLEG): container finished" podID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerID="38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2" exitCode=0 Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.426100 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c9vsk" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.426100 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerDied","Data":"38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2"} Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.426346 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c9vsk" event={"ID":"3894718b-9c8e-4256-8f71-9bfe377bac10","Type":"ContainerDied","Data":"93b174f0293e82c407f67c773437c9a34c527eaf960677aca48663c04488707d"} Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.426371 4775 scope.go:117] "RemoveContainer" containerID="38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.454212 4775 scope.go:117] "RemoveContainer" containerID="11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.463463 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c9vsk"] Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.476795 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c9vsk"] Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.477363 4775 scope.go:117] "RemoveContainer" containerID="cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.526258 4775 scope.go:117] "RemoveContainer" containerID="38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2" Nov 26 07:13:20 crc kubenswrapper[4775]: E1126 07:13:20.526642 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2\": container with ID starting with 38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2 not found: ID does not exist" containerID="38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.526692 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2"} err="failed to get container status \"38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2\": rpc error: code = NotFound desc = could not find container \"38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2\": container with ID starting with 38b889acedb2438032b91b82de5e5b457d7a98b9bbbfc8df429c188c0b291ae2 not found: ID does not exist" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.526735 4775 scope.go:117] "RemoveContainer" containerID="11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035" Nov 26 07:13:20 crc kubenswrapper[4775]: E1126 07:13:20.527065 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035\": container with ID starting with 11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035 not found: ID does not exist" containerID="11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.527094 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035"} err="failed to get container status \"11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035\": rpc error: code = NotFound desc = could not find container \"11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035\": container with ID starting with 11184793d2a1c6ec0f973f2ebaba0ddd557a136c4a707e3e3aa5cfac00b96035 not found: ID does not exist" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.527110 4775 scope.go:117] "RemoveContainer" containerID="cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889" Nov 26 07:13:20 crc kubenswrapper[4775]: E1126 07:13:20.527536 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889\": container with ID starting with cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889 not found: ID does not exist" containerID="cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889" Nov 26 07:13:20 crc kubenswrapper[4775]: I1126 07:13:20.527573 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889"} err="failed to get container status \"cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889\": rpc error: code = NotFound desc = could not find container \"cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889\": container with ID starting with cc1b6c624f1eaeaa06dced5e0640f2a57231382d950ec1a65f0f18620c6fd889 not found: ID does not exist" Nov 26 07:13:21 crc kubenswrapper[4775]: I1126 07:13:21.339866 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" path="/var/lib/kubelet/pods/3894718b-9c8e-4256-8f71-9bfe377bac10/volumes" Nov 26 07:13:28 crc kubenswrapper[4775]: I1126 07:13:28.328139 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:13:28 crc kubenswrapper[4775]: E1126 07:13:28.329259 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:13:40 crc kubenswrapper[4775]: I1126 07:13:40.327568 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:13:40 crc kubenswrapper[4775]: E1126 07:13:40.328406 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:13:52 crc kubenswrapper[4775]: I1126 07:13:52.328145 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:13:52 crc kubenswrapper[4775]: E1126 07:13:52.329313 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:14:07 crc kubenswrapper[4775]: I1126 07:14:07.337045 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:14:07 crc kubenswrapper[4775]: E1126 07:14:07.337641 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:14:21 crc kubenswrapper[4775]: I1126 07:14:21.328608 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:14:21 crc kubenswrapper[4775]: E1126 07:14:21.329567 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:14:36 crc kubenswrapper[4775]: I1126 07:14:36.329250 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:14:36 crc kubenswrapper[4775]: E1126 07:14:36.331130 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:14:49 crc kubenswrapper[4775]: I1126 07:14:49.329095 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:14:49 crc kubenswrapper[4775]: E1126 07:14:49.330152 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.151934 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls"] Nov 26 07:15:00 crc kubenswrapper[4775]: E1126 07:15:00.152759 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="extract-utilities" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.152777 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="extract-utilities" Nov 26 07:15:00 crc kubenswrapper[4775]: E1126 07:15:00.152825 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="extract-content" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.152831 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="extract-content" Nov 26 07:15:00 crc kubenswrapper[4775]: E1126 07:15:00.152849 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="registry-server" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.152855 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="registry-server" Nov 26 07:15:00 crc kubenswrapper[4775]: E1126 07:15:00.152879 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="extract-utilities" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.152884 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="extract-utilities" Nov 26 07:15:00 crc kubenswrapper[4775]: E1126 07:15:00.152900 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="registry-server" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.152905 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="registry-server" Nov 26 07:15:00 crc kubenswrapper[4775]: E1126 07:15:00.152916 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="extract-content" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.152922 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="extract-content" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.153124 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a67f4c-f3c9-4ef8-8224-5e1ab3944f99" containerName="registry-server" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.153136 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3894718b-9c8e-4256-8f71-9bfe377bac10" containerName="registry-server" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.156441 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.159359 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.159546 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.174757 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls"] Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.257735 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f339d95-5bc0-423b-ac01-3792460031f0-secret-volume\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.257888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jshh4\" (UniqueName: \"kubernetes.io/projected/2f339d95-5bc0-423b-ac01-3792460031f0-kube-api-access-jshh4\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.257962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f339d95-5bc0-423b-ac01-3792460031f0-config-volume\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.360151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f339d95-5bc0-423b-ac01-3792460031f0-secret-volume\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.360273 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jshh4\" (UniqueName: \"kubernetes.io/projected/2f339d95-5bc0-423b-ac01-3792460031f0-kube-api-access-jshh4\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.360328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f339d95-5bc0-423b-ac01-3792460031f0-config-volume\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.361405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f339d95-5bc0-423b-ac01-3792460031f0-config-volume\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.373328 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f339d95-5bc0-423b-ac01-3792460031f0-secret-volume\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.386699 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jshh4\" (UniqueName: \"kubernetes.io/projected/2f339d95-5bc0-423b-ac01-3792460031f0-kube-api-access-jshh4\") pod \"collect-profiles-29402355-l68ls\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:00 crc kubenswrapper[4775]: I1126 07:15:00.491407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:01 crc kubenswrapper[4775]: I1126 07:15:01.012571 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls"] Nov 26 07:15:01 crc kubenswrapper[4775]: I1126 07:15:01.359475 4775 generic.go:334] "Generic (PLEG): container finished" podID="2f339d95-5bc0-423b-ac01-3792460031f0" containerID="49ca98605d1823b782decb9442c06b851ab7154f4b428b618845bcb771bebc7c" exitCode=0 Nov 26 07:15:01 crc kubenswrapper[4775]: I1126 07:15:01.359519 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" event={"ID":"2f339d95-5bc0-423b-ac01-3792460031f0","Type":"ContainerDied","Data":"49ca98605d1823b782decb9442c06b851ab7154f4b428b618845bcb771bebc7c"} Nov 26 07:15:01 crc kubenswrapper[4775]: I1126 07:15:01.359545 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" event={"ID":"2f339d95-5bc0-423b-ac01-3792460031f0","Type":"ContainerStarted","Data":"f92bd213525db5f46bc204fc3b98cac3d5455d90795aac03bf92b646df207020"} Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.775605 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.803438 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f339d95-5bc0-423b-ac01-3792460031f0-config-volume\") pod \"2f339d95-5bc0-423b-ac01-3792460031f0\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.803751 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f339d95-5bc0-423b-ac01-3792460031f0-secret-volume\") pod \"2f339d95-5bc0-423b-ac01-3792460031f0\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.803826 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jshh4\" (UniqueName: \"kubernetes.io/projected/2f339d95-5bc0-423b-ac01-3792460031f0-kube-api-access-jshh4\") pod \"2f339d95-5bc0-423b-ac01-3792460031f0\" (UID: \"2f339d95-5bc0-423b-ac01-3792460031f0\") " Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.804317 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f339d95-5bc0-423b-ac01-3792460031f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f339d95-5bc0-423b-ac01-3792460031f0" (UID: "2f339d95-5bc0-423b-ac01-3792460031f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.810661 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f339d95-5bc0-423b-ac01-3792460031f0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.811933 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f339d95-5bc0-423b-ac01-3792460031f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f339d95-5bc0-423b-ac01-3792460031f0" (UID: "2f339d95-5bc0-423b-ac01-3792460031f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.811975 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f339d95-5bc0-423b-ac01-3792460031f0-kube-api-access-jshh4" (OuterVolumeSpecName: "kube-api-access-jshh4") pod "2f339d95-5bc0-423b-ac01-3792460031f0" (UID: "2f339d95-5bc0-423b-ac01-3792460031f0"). InnerVolumeSpecName "kube-api-access-jshh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.912908 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f339d95-5bc0-423b-ac01-3792460031f0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 07:15:02 crc kubenswrapper[4775]: I1126 07:15:02.912952 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jshh4\" (UniqueName: \"kubernetes.io/projected/2f339d95-5bc0-423b-ac01-3792460031f0-kube-api-access-jshh4\") on node \"crc\" DevicePath \"\"" Nov 26 07:15:03 crc kubenswrapper[4775]: I1126 07:15:03.327857 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:15:03 crc kubenswrapper[4775]: E1126 07:15:03.328382 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:15:03 crc kubenswrapper[4775]: I1126 07:15:03.385909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" event={"ID":"2f339d95-5bc0-423b-ac01-3792460031f0","Type":"ContainerDied","Data":"f92bd213525db5f46bc204fc3b98cac3d5455d90795aac03bf92b646df207020"} Nov 26 07:15:03 crc kubenswrapper[4775]: I1126 07:15:03.385949 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402355-l68ls" Nov 26 07:15:03 crc kubenswrapper[4775]: I1126 07:15:03.385965 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f92bd213525db5f46bc204fc3b98cac3d5455d90795aac03bf92b646df207020" Nov 26 07:15:03 crc kubenswrapper[4775]: E1126 07:15:03.565981 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f339d95_5bc0_423b_ac01_3792460031f0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f339d95_5bc0_423b_ac01_3792460031f0.slice/crio-f92bd213525db5f46bc204fc3b98cac3d5455d90795aac03bf92b646df207020\": RecentStats: unable to find data in memory cache]" Nov 26 07:15:03 crc kubenswrapper[4775]: I1126 07:15:03.838442 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4"] Nov 26 07:15:03 crc kubenswrapper[4775]: I1126 07:15:03.846877 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402310-mhmw4"] Nov 26 07:15:05 crc kubenswrapper[4775]: I1126 07:15:05.346584 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c54c2235-f554-4eda-bae7-b4fa7b68891e" path="/var/lib/kubelet/pods/c54c2235-f554-4eda-bae7-b4fa7b68891e/volumes" Nov 26 07:15:14 crc kubenswrapper[4775]: I1126 07:15:14.328213 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:15:14 crc kubenswrapper[4775]: E1126 07:15:14.329281 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:15:22 crc kubenswrapper[4775]: I1126 07:15:22.592856 4775 scope.go:117] "RemoveContainer" containerID="a8945445912302fd8e3eba0af31209a72a6a00ed0850130cd144f9c71fc83233" Nov 26 07:15:28 crc kubenswrapper[4775]: I1126 07:15:28.328432 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:15:28 crc kubenswrapper[4775]: E1126 07:15:28.329357 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:15:39 crc kubenswrapper[4775]: I1126 07:15:39.328107 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:15:39 crc kubenswrapper[4775]: E1126 07:15:39.328984 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:15:54 crc kubenswrapper[4775]: I1126 07:15:54.327688 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:15:54 crc kubenswrapper[4775]: E1126 07:15:54.328441 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:16:08 crc kubenswrapper[4775]: I1126 07:16:08.327907 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:16:08 crc kubenswrapper[4775]: E1126 07:16:08.328839 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:16:22 crc kubenswrapper[4775]: I1126 07:16:22.327419 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:16:22 crc kubenswrapper[4775]: E1126 07:16:22.328064 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:16:37 crc kubenswrapper[4775]: I1126 07:16:37.336873 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:16:37 crc kubenswrapper[4775]: E1126 07:16:37.337772 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:16:49 crc kubenswrapper[4775]: I1126 07:16:49.327513 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:16:49 crc kubenswrapper[4775]: E1126 07:16:49.328374 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:17:00 crc kubenswrapper[4775]: I1126 07:17:00.328578 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:17:00 crc kubenswrapper[4775]: E1126 07:17:00.329367 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:17:11 crc kubenswrapper[4775]: I1126 07:17:11.328171 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:17:11 crc kubenswrapper[4775]: E1126 07:17:11.328882 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:17:24 crc kubenswrapper[4775]: I1126 07:17:24.328009 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:17:24 crc kubenswrapper[4775]: E1126 07:17:24.328764 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:17:39 crc kubenswrapper[4775]: I1126 07:17:39.327762 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:17:39 crc kubenswrapper[4775]: E1126 07:17:39.328544 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:17:52 crc kubenswrapper[4775]: I1126 07:17:52.328033 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:17:53 crc kubenswrapper[4775]: I1126 07:17:53.204959 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"cee8e65f1faefe50945972b8e3690b32b04bca237ff59c2b933c2dacd5b41745"} Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.828835 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tx78g"] Nov 26 07:20:08 crc kubenswrapper[4775]: E1126 07:20:08.829923 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f339d95-5bc0-423b-ac01-3792460031f0" containerName="collect-profiles" Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.829943 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f339d95-5bc0-423b-ac01-3792460031f0" containerName="collect-profiles" Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.830229 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f339d95-5bc0-423b-ac01-3792460031f0" containerName="collect-profiles" Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.832285 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.844601 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tx78g"] Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.997287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-catalog-content\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.997365 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbgxc\" (UniqueName: \"kubernetes.io/projected/895c1483-72a6-4296-85fc-32acd5ddde91-kube-api-access-wbgxc\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:08 crc kubenswrapper[4775]: I1126 07:20:08.997434 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-utilities\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.099391 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-catalog-content\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.099460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbgxc\" (UniqueName: \"kubernetes.io/projected/895c1483-72a6-4296-85fc-32acd5ddde91-kube-api-access-wbgxc\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.099513 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-utilities\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.100024 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-utilities\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.100023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-catalog-content\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.125612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbgxc\" (UniqueName: \"kubernetes.io/projected/895c1483-72a6-4296-85fc-32acd5ddde91-kube-api-access-wbgxc\") pod \"certified-operators-tx78g\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.156745 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:09 crc kubenswrapper[4775]: I1126 07:20:09.741637 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tx78g"] Nov 26 07:20:10 crc kubenswrapper[4775]: I1126 07:20:10.373291 4775 generic.go:334] "Generic (PLEG): container finished" podID="895c1483-72a6-4296-85fc-32acd5ddde91" containerID="4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da" exitCode=0 Nov 26 07:20:10 crc kubenswrapper[4775]: I1126 07:20:10.373332 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tx78g" event={"ID":"895c1483-72a6-4296-85fc-32acd5ddde91","Type":"ContainerDied","Data":"4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da"} Nov 26 07:20:10 crc kubenswrapper[4775]: I1126 07:20:10.373356 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tx78g" event={"ID":"895c1483-72a6-4296-85fc-32acd5ddde91","Type":"ContainerStarted","Data":"71ac3bf4d72a4eea97b812b019f6d2a77794afad9bfa76a89650164bcbbeb9c4"} Nov 26 07:20:10 crc kubenswrapper[4775]: I1126 07:20:10.375934 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 07:20:11 crc kubenswrapper[4775]: I1126 07:20:11.384118 4775 generic.go:334] "Generic (PLEG): container finished" podID="895c1483-72a6-4296-85fc-32acd5ddde91" containerID="ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a" exitCode=0 Nov 26 07:20:11 crc kubenswrapper[4775]: I1126 07:20:11.384185 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tx78g" event={"ID":"895c1483-72a6-4296-85fc-32acd5ddde91","Type":"ContainerDied","Data":"ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a"} Nov 26 07:20:12 crc kubenswrapper[4775]: I1126 07:20:12.395866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tx78g" event={"ID":"895c1483-72a6-4296-85fc-32acd5ddde91","Type":"ContainerStarted","Data":"07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe"} Nov 26 07:20:19 crc kubenswrapper[4775]: I1126 07:20:19.157647 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:19 crc kubenswrapper[4775]: I1126 07:20:19.158261 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:19 crc kubenswrapper[4775]: I1126 07:20:19.231665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:19 crc kubenswrapper[4775]: I1126 07:20:19.262131 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tx78g" podStartSLOduration=9.839267996 podStartE2EDuration="11.262108812s" podCreationTimestamp="2025-11-26 07:20:08 +0000 UTC" firstStartedPulling="2025-11-26 07:20:10.375676822 +0000 UTC m=+3293.736980774" lastFinishedPulling="2025-11-26 07:20:11.798517638 +0000 UTC m=+3295.159821590" observedRunningTime="2025-11-26 07:20:12.427911598 +0000 UTC m=+3295.789215550" watchObservedRunningTime="2025-11-26 07:20:19.262108812 +0000 UTC m=+3302.623412784" Nov 26 07:20:19 crc kubenswrapper[4775]: I1126 07:20:19.499993 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:19 crc kubenswrapper[4775]: I1126 07:20:19.554789 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tx78g"] Nov 26 07:20:21 crc kubenswrapper[4775]: I1126 07:20:21.419963 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:20:21 crc kubenswrapper[4775]: I1126 07:20:21.420376 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:20:21 crc kubenswrapper[4775]: I1126 07:20:21.475319 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tx78g" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="registry-server" containerID="cri-o://07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe" gracePeriod=2 Nov 26 07:20:21 crc kubenswrapper[4775]: I1126 07:20:21.986778 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.146672 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbgxc\" (UniqueName: \"kubernetes.io/projected/895c1483-72a6-4296-85fc-32acd5ddde91-kube-api-access-wbgxc\") pod \"895c1483-72a6-4296-85fc-32acd5ddde91\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.147344 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-catalog-content\") pod \"895c1483-72a6-4296-85fc-32acd5ddde91\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.147522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-utilities\") pod \"895c1483-72a6-4296-85fc-32acd5ddde91\" (UID: \"895c1483-72a6-4296-85fc-32acd5ddde91\") " Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.148299 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-utilities" (OuterVolumeSpecName: "utilities") pod "895c1483-72a6-4296-85fc-32acd5ddde91" (UID: "895c1483-72a6-4296-85fc-32acd5ddde91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.153634 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/895c1483-72a6-4296-85fc-32acd5ddde91-kube-api-access-wbgxc" (OuterVolumeSpecName: "kube-api-access-wbgxc") pod "895c1483-72a6-4296-85fc-32acd5ddde91" (UID: "895c1483-72a6-4296-85fc-32acd5ddde91"). InnerVolumeSpecName "kube-api-access-wbgxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.202788 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "895c1483-72a6-4296-85fc-32acd5ddde91" (UID: "895c1483-72a6-4296-85fc-32acd5ddde91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.250011 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.250047 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/895c1483-72a6-4296-85fc-32acd5ddde91-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.250062 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbgxc\" (UniqueName: \"kubernetes.io/projected/895c1483-72a6-4296-85fc-32acd5ddde91-kube-api-access-wbgxc\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.488447 4775 generic.go:334] "Generic (PLEG): container finished" podID="895c1483-72a6-4296-85fc-32acd5ddde91" containerID="07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe" exitCode=0 Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.488502 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tx78g" event={"ID":"895c1483-72a6-4296-85fc-32acd5ddde91","Type":"ContainerDied","Data":"07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe"} Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.488528 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tx78g" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.488557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tx78g" event={"ID":"895c1483-72a6-4296-85fc-32acd5ddde91","Type":"ContainerDied","Data":"71ac3bf4d72a4eea97b812b019f6d2a77794afad9bfa76a89650164bcbbeb9c4"} Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.488584 4775 scope.go:117] "RemoveContainer" containerID="07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.526117 4775 scope.go:117] "RemoveContainer" containerID="ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.530521 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tx78g"] Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.539731 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tx78g"] Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.552408 4775 scope.go:117] "RemoveContainer" containerID="4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.603992 4775 scope.go:117] "RemoveContainer" containerID="07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe" Nov 26 07:20:22 crc kubenswrapper[4775]: E1126 07:20:22.604751 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe\": container with ID starting with 07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe not found: ID does not exist" containerID="07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.604814 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe"} err="failed to get container status \"07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe\": rpc error: code = NotFound desc = could not find container \"07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe\": container with ID starting with 07b15600da370d452535e6bc6decef5fd1a6167387edb36322a552319046b6fe not found: ID does not exist" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.604853 4775 scope.go:117] "RemoveContainer" containerID="ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a" Nov 26 07:20:22 crc kubenswrapper[4775]: E1126 07:20:22.605405 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a\": container with ID starting with ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a not found: ID does not exist" containerID="ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.605438 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a"} err="failed to get container status \"ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a\": rpc error: code = NotFound desc = could not find container \"ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a\": container with ID starting with ad256d7486c68c012e2bf536615876da5a424f04de5354df4bc73fc1783d293a not found: ID does not exist" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.605462 4775 scope.go:117] "RemoveContainer" containerID="4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da" Nov 26 07:20:22 crc kubenswrapper[4775]: E1126 07:20:22.605787 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da\": container with ID starting with 4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da not found: ID does not exist" containerID="4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da" Nov 26 07:20:22 crc kubenswrapper[4775]: I1126 07:20:22.605810 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da"} err="failed to get container status \"4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da\": rpc error: code = NotFound desc = could not find container \"4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da\": container with ID starting with 4c68e0560aa10ad39f1457b4f5ccfa3ec26815061dc699a39160ef1dbf86b9da not found: ID does not exist" Nov 26 07:20:23 crc kubenswrapper[4775]: I1126 07:20:23.343343 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" path="/var/lib/kubelet/pods/895c1483-72a6-4296-85fc-32acd5ddde91/volumes" Nov 26 07:20:23 crc kubenswrapper[4775]: I1126 07:20:23.503148 4775 generic.go:334] "Generic (PLEG): container finished" podID="da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" containerID="3e57123e18cda991cc838d510f81306094d91df222c84e96f42e9d4be423739e" exitCode=0 Nov 26 07:20:23 crc kubenswrapper[4775]: I1126 07:20:23.503202 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5","Type":"ContainerDied","Data":"3e57123e18cda991cc838d510f81306094d91df222c84e96f42e9d4be423739e"} Nov 26 07:20:24 crc kubenswrapper[4775]: I1126 07:20:24.849660 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002056 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msvkm\" (UniqueName: \"kubernetes.io/projected/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-kube-api-access-msvkm\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ssh-key\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002471 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-temporary\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002561 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002673 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ca-certs\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002772 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-workdir\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002853 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config-secret\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.002932 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.003032 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-config-data\") pod \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\" (UID: \"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5\") " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.003539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-config-data" (OuterVolumeSpecName: "config-data") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.003549 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.008941 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.009037 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-kube-api-access-msvkm" (OuterVolumeSpecName: "kube-api-access-msvkm") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "kube-api-access-msvkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.010682 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.033531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.042496 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.051843 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.094978 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" (UID: "da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105304 4775 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105463 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105524 4775 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105580 4775 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105657 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105749 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.105815 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.106335 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msvkm\" (UniqueName: \"kubernetes.io/projected/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-kube-api-access-msvkm\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.106398 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.121882 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.208119 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.532622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5","Type":"ContainerDied","Data":"400ef64c5ebb828ab92a46428f5aff3daccc116ad55b31d079d9692f2972ce63"} Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.532666 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="400ef64c5ebb828ab92a46428f5aff3daccc116ad55b31d079d9692f2972ce63" Nov 26 07:20:25 crc kubenswrapper[4775]: I1126 07:20:25.532698 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.017640 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 26 07:20:28 crc kubenswrapper[4775]: E1126 07:20:28.018548 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" containerName="tempest-tests-tempest-tests-runner" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.018570 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" containerName="tempest-tests-tempest-tests-runner" Nov 26 07:20:28 crc kubenswrapper[4775]: E1126 07:20:28.018588 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="extract-content" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.018598 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="extract-content" Nov 26 07:20:28 crc kubenswrapper[4775]: E1126 07:20:28.018617 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="registry-server" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.018626 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="registry-server" Nov 26 07:20:28 crc kubenswrapper[4775]: E1126 07:20:28.018657 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="extract-utilities" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.018667 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="extract-utilities" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.018996 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="895c1483-72a6-4296-85fc-32acd5ddde91" containerName="registry-server" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.019028 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5" containerName="tempest-tests-tempest-tests-runner" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.019985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.023567 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5p94c" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.043898 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.159614 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.159673 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-652cm\" (UniqueName: \"kubernetes.io/projected/7075bbd0-afa7-4430-b2b9-ee07eac73b2a-kube-api-access-652cm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.262012 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.262075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-652cm\" (UniqueName: \"kubernetes.io/projected/7075bbd0-afa7-4430-b2b9-ee07eac73b2a-kube-api-access-652cm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.262479 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.313497 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-652cm\" (UniqueName: \"kubernetes.io/projected/7075bbd0-afa7-4430-b2b9-ee07eac73b2a-kube-api-access-652cm\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.325770 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"7075bbd0-afa7-4430-b2b9-ee07eac73b2a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.361073 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 26 07:20:28 crc kubenswrapper[4775]: I1126 07:20:28.876105 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 26 07:20:29 crc kubenswrapper[4775]: I1126 07:20:29.574994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7075bbd0-afa7-4430-b2b9-ee07eac73b2a","Type":"ContainerStarted","Data":"378b4f2f9752f96caad115a630d29870dcdbe69cdf731249d8c9d667b0701d7f"} Nov 26 07:20:30 crc kubenswrapper[4775]: I1126 07:20:30.589000 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"7075bbd0-afa7-4430-b2b9-ee07eac73b2a","Type":"ContainerStarted","Data":"33b2802a87372d2e5fe8a2466889ba4ec4e38d0b7de6a2961bc526a3597f8958"} Nov 26 07:20:30 crc kubenswrapper[4775]: I1126 07:20:30.616233 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.7646184099999997 podStartE2EDuration="3.616202769s" podCreationTimestamp="2025-11-26 07:20:27 +0000 UTC" firstStartedPulling="2025-11-26 07:20:28.880220715 +0000 UTC m=+3312.241524667" lastFinishedPulling="2025-11-26 07:20:29.731805074 +0000 UTC m=+3313.093109026" observedRunningTime="2025-11-26 07:20:30.607103026 +0000 UTC m=+3313.968407018" watchObservedRunningTime="2025-11-26 07:20:30.616202769 +0000 UTC m=+3313.977506751" Nov 26 07:20:51 crc kubenswrapper[4775]: I1126 07:20:51.419661 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:20:51 crc kubenswrapper[4775]: I1126 07:20:51.420155 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.338179 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48sft/must-gather-ljm82"] Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.340232 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.342879 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-48sft"/"openshift-service-ca.crt" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.351181 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-48sft"/"kube-root-ca.crt" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.372478 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-48sft/must-gather-ljm82"] Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.439012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad9f4c47-c3af-444f-9b15-48b40663f073-must-gather-output\") pod \"must-gather-ljm82\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.439137 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nz28\" (UniqueName: \"kubernetes.io/projected/ad9f4c47-c3af-444f-9b15-48b40663f073-kube-api-access-8nz28\") pod \"must-gather-ljm82\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.541067 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nz28\" (UniqueName: \"kubernetes.io/projected/ad9f4c47-c3af-444f-9b15-48b40663f073-kube-api-access-8nz28\") pod \"must-gather-ljm82\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.541300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad9f4c47-c3af-444f-9b15-48b40663f073-must-gather-output\") pod \"must-gather-ljm82\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.541841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad9f4c47-c3af-444f-9b15-48b40663f073-must-gather-output\") pod \"must-gather-ljm82\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.563500 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nz28\" (UniqueName: \"kubernetes.io/projected/ad9f4c47-c3af-444f-9b15-48b40663f073-kube-api-access-8nz28\") pod \"must-gather-ljm82\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:53 crc kubenswrapper[4775]: I1126 07:20:53.669276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:20:54 crc kubenswrapper[4775]: I1126 07:20:54.127353 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-48sft/must-gather-ljm82"] Nov 26 07:20:54 crc kubenswrapper[4775]: I1126 07:20:54.820401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/must-gather-ljm82" event={"ID":"ad9f4c47-c3af-444f-9b15-48b40663f073","Type":"ContainerStarted","Data":"521e13e0c422f9e3343d2dd6435773df730c9f23694344f3820f8259c2308a0f"} Nov 26 07:20:58 crc kubenswrapper[4775]: I1126 07:20:58.857627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/must-gather-ljm82" event={"ID":"ad9f4c47-c3af-444f-9b15-48b40663f073","Type":"ContainerStarted","Data":"ffb56d09d07259fc8c523edb26e695b5909d0ad0b85b5d8d9a74991a8f99cd2d"} Nov 26 07:20:58 crc kubenswrapper[4775]: I1126 07:20:58.858208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/must-gather-ljm82" event={"ID":"ad9f4c47-c3af-444f-9b15-48b40663f073","Type":"ContainerStarted","Data":"3354087f787034c3409087b94b6d623d122f76079d4afe12d9d82fa27a3f8c1f"} Nov 26 07:20:58 crc kubenswrapper[4775]: I1126 07:20:58.882864 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-48sft/must-gather-ljm82" podStartSLOduration=1.759617588 podStartE2EDuration="5.882839581s" podCreationTimestamp="2025-11-26 07:20:53 +0000 UTC" firstStartedPulling="2025-11-26 07:20:54.133481825 +0000 UTC m=+3337.494785777" lastFinishedPulling="2025-11-26 07:20:58.256703818 +0000 UTC m=+3341.618007770" observedRunningTime="2025-11-26 07:20:58.87530222 +0000 UTC m=+3342.236606172" watchObservedRunningTime="2025-11-26 07:20:58.882839581 +0000 UTC m=+3342.244143533" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.222059 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48sft/crc-debug-wsxn6"] Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.223786 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.226173 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-48sft"/"default-dockercfg-mlfmh" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.412836 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e00d008-1694-4b74-930b-7e58b949a5c6-host\") pod \"crc-debug-wsxn6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.412888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k875v\" (UniqueName: \"kubernetes.io/projected/2e00d008-1694-4b74-930b-7e58b949a5c6-kube-api-access-k875v\") pod \"crc-debug-wsxn6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.514418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e00d008-1694-4b74-930b-7e58b949a5c6-host\") pod \"crc-debug-wsxn6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.514462 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k875v\" (UniqueName: \"kubernetes.io/projected/2e00d008-1694-4b74-930b-7e58b949a5c6-kube-api-access-k875v\") pod \"crc-debug-wsxn6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.514543 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e00d008-1694-4b74-930b-7e58b949a5c6-host\") pod \"crc-debug-wsxn6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.541502 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k875v\" (UniqueName: \"kubernetes.io/projected/2e00d008-1694-4b74-930b-7e58b949a5c6-kube-api-access-k875v\") pod \"crc-debug-wsxn6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.550165 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:02 crc kubenswrapper[4775]: I1126 07:21:02.890559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-wsxn6" event={"ID":"2e00d008-1694-4b74-930b-7e58b949a5c6","Type":"ContainerStarted","Data":"1dbd43cbaf8b33b147430bcf632e3c549128c986027c31496f0eff8bb4cdcecc"} Nov 26 07:21:15 crc kubenswrapper[4775]: I1126 07:21:15.008974 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-wsxn6" event={"ID":"2e00d008-1694-4b74-930b-7e58b949a5c6","Type":"ContainerStarted","Data":"7c51fca7ddf0ab94168303f8f14cb557b25e00703eba689c349b627668c4edaf"} Nov 26 07:21:15 crc kubenswrapper[4775]: I1126 07:21:15.027704 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-48sft/crc-debug-wsxn6" podStartSLOduration=1.51335534 podStartE2EDuration="13.027687049s" podCreationTimestamp="2025-11-26 07:21:02 +0000 UTC" firstStartedPulling="2025-11-26 07:21:02.583288055 +0000 UTC m=+3345.944592007" lastFinishedPulling="2025-11-26 07:21:14.097619764 +0000 UTC m=+3357.458923716" observedRunningTime="2025-11-26 07:21:15.023735444 +0000 UTC m=+3358.385039406" watchObservedRunningTime="2025-11-26 07:21:15.027687049 +0000 UTC m=+3358.388990991" Nov 26 07:21:21 crc kubenswrapper[4775]: I1126 07:21:21.419371 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:21:21 crc kubenswrapper[4775]: I1126 07:21:21.420097 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:21:21 crc kubenswrapper[4775]: I1126 07:21:21.420166 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:21:21 crc kubenswrapper[4775]: I1126 07:21:21.421133 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cee8e65f1faefe50945972b8e3690b32b04bca237ff59c2b933c2dacd5b41745"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:21:21 crc kubenswrapper[4775]: I1126 07:21:21.421201 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://cee8e65f1faefe50945972b8e3690b32b04bca237ff59c2b933c2dacd5b41745" gracePeriod=600 Nov 26 07:21:22 crc kubenswrapper[4775]: I1126 07:21:22.095463 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="cee8e65f1faefe50945972b8e3690b32b04bca237ff59c2b933c2dacd5b41745" exitCode=0 Nov 26 07:21:22 crc kubenswrapper[4775]: I1126 07:21:22.095593 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"cee8e65f1faefe50945972b8e3690b32b04bca237ff59c2b933c2dacd5b41745"} Nov 26 07:21:22 crc kubenswrapper[4775]: I1126 07:21:22.095836 4775 scope.go:117] "RemoveContainer" containerID="252da7c937f7bfa3cec3ea0c276bc93b358bb56d1ad5ebe295ef3db97cc789b5" Nov 26 07:21:27 crc kubenswrapper[4775]: I1126 07:21:27.156811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531"} Nov 26 07:21:56 crc kubenswrapper[4775]: I1126 07:21:56.412617 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e00d008-1694-4b74-930b-7e58b949a5c6" containerID="7c51fca7ddf0ab94168303f8f14cb557b25e00703eba689c349b627668c4edaf" exitCode=0 Nov 26 07:21:56 crc kubenswrapper[4775]: I1126 07:21:56.412740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-wsxn6" event={"ID":"2e00d008-1694-4b74-930b-7e58b949a5c6","Type":"ContainerDied","Data":"7c51fca7ddf0ab94168303f8f14cb557b25e00703eba689c349b627668c4edaf"} Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.545658 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.581010 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48sft/crc-debug-wsxn6"] Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.590533 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48sft/crc-debug-wsxn6"] Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.699393 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e00d008-1694-4b74-930b-7e58b949a5c6-host\") pod \"2e00d008-1694-4b74-930b-7e58b949a5c6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.699550 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k875v\" (UniqueName: \"kubernetes.io/projected/2e00d008-1694-4b74-930b-7e58b949a5c6-kube-api-access-k875v\") pod \"2e00d008-1694-4b74-930b-7e58b949a5c6\" (UID: \"2e00d008-1694-4b74-930b-7e58b949a5c6\") " Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.699874 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e00d008-1694-4b74-930b-7e58b949a5c6-host" (OuterVolumeSpecName: "host") pod "2e00d008-1694-4b74-930b-7e58b949a5c6" (UID: "2e00d008-1694-4b74-930b-7e58b949a5c6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.700327 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2e00d008-1694-4b74-930b-7e58b949a5c6-host\") on node \"crc\" DevicePath \"\"" Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.704937 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e00d008-1694-4b74-930b-7e58b949a5c6-kube-api-access-k875v" (OuterVolumeSpecName: "kube-api-access-k875v") pod "2e00d008-1694-4b74-930b-7e58b949a5c6" (UID: "2e00d008-1694-4b74-930b-7e58b949a5c6"). InnerVolumeSpecName "kube-api-access-k875v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:21:57 crc kubenswrapper[4775]: I1126 07:21:57.802004 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k875v\" (UniqueName: \"kubernetes.io/projected/2e00d008-1694-4b74-930b-7e58b949a5c6-kube-api-access-k875v\") on node \"crc\" DevicePath \"\"" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.432278 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dbd43cbaf8b33b147430bcf632e3c549128c986027c31496f0eff8bb4cdcecc" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.432332 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-wsxn6" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.736751 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48sft/crc-debug-p24tq"] Nov 26 07:21:58 crc kubenswrapper[4775]: E1126 07:21:58.737466 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e00d008-1694-4b74-930b-7e58b949a5c6" containerName="container-00" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.737481 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e00d008-1694-4b74-930b-7e58b949a5c6" containerName="container-00" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.737664 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e00d008-1694-4b74-930b-7e58b949a5c6" containerName="container-00" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.738253 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.740351 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-48sft"/"default-dockercfg-mlfmh" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.923185 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxxds\" (UniqueName: \"kubernetes.io/projected/7fc49afc-b637-4c08-8318-8f9ca921569c-kube-api-access-rxxds\") pod \"crc-debug-p24tq\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:58 crc kubenswrapper[4775]: I1126 07:21:58.923593 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc49afc-b637-4c08-8318-8f9ca921569c-host\") pod \"crc-debug-p24tq\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.026013 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxxds\" (UniqueName: \"kubernetes.io/projected/7fc49afc-b637-4c08-8318-8f9ca921569c-kube-api-access-rxxds\") pod \"crc-debug-p24tq\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.026129 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc49afc-b637-4c08-8318-8f9ca921569c-host\") pod \"crc-debug-p24tq\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.026385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc49afc-b637-4c08-8318-8f9ca921569c-host\") pod \"crc-debug-p24tq\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.045507 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxxds\" (UniqueName: \"kubernetes.io/projected/7fc49afc-b637-4c08-8318-8f9ca921569c-kube-api-access-rxxds\") pod \"crc-debug-p24tq\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.064136 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.338308 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e00d008-1694-4b74-930b-7e58b949a5c6" path="/var/lib/kubelet/pods/2e00d008-1694-4b74-930b-7e58b949a5c6/volumes" Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.443670 4775 generic.go:334] "Generic (PLEG): container finished" podID="7fc49afc-b637-4c08-8318-8f9ca921569c" containerID="da28ac7b15ad7c77dadcdc06cce1e50ca0154de6c25173000f3270c06bee7f08" exitCode=0 Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.443746 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-p24tq" event={"ID":"7fc49afc-b637-4c08-8318-8f9ca921569c","Type":"ContainerDied","Data":"da28ac7b15ad7c77dadcdc06cce1e50ca0154de6c25173000f3270c06bee7f08"} Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.443933 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-p24tq" event={"ID":"7fc49afc-b637-4c08-8318-8f9ca921569c","Type":"ContainerStarted","Data":"9087e174c5653640cce4ce1eeade3b5b75655cc2e0d4f75805359f5fbea1669b"} Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.923384 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48sft/crc-debug-p24tq"] Nov 26 07:21:59 crc kubenswrapper[4775]: I1126 07:21:59.931952 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48sft/crc-debug-p24tq"] Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.581405 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.763371 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc49afc-b637-4c08-8318-8f9ca921569c-host\") pod \"7fc49afc-b637-4c08-8318-8f9ca921569c\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.763534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxxds\" (UniqueName: \"kubernetes.io/projected/7fc49afc-b637-4c08-8318-8f9ca921569c-kube-api-access-rxxds\") pod \"7fc49afc-b637-4c08-8318-8f9ca921569c\" (UID: \"7fc49afc-b637-4c08-8318-8f9ca921569c\") " Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.763620 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fc49afc-b637-4c08-8318-8f9ca921569c-host" (OuterVolumeSpecName: "host") pod "7fc49afc-b637-4c08-8318-8f9ca921569c" (UID: "7fc49afc-b637-4c08-8318-8f9ca921569c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.764224 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc49afc-b637-4c08-8318-8f9ca921569c-host\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.770298 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc49afc-b637-4c08-8318-8f9ca921569c-kube-api-access-rxxds" (OuterVolumeSpecName: "kube-api-access-rxxds") pod "7fc49afc-b637-4c08-8318-8f9ca921569c" (UID: "7fc49afc-b637-4c08-8318-8f9ca921569c"). InnerVolumeSpecName "kube-api-access-rxxds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:22:00 crc kubenswrapper[4775]: I1126 07:22:00.865915 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxxds\" (UniqueName: \"kubernetes.io/projected/7fc49afc-b637-4c08-8318-8f9ca921569c-kube-api-access-rxxds\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.102224 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-48sft/crc-debug-7mlwk"] Nov 26 07:22:01 crc kubenswrapper[4775]: E1126 07:22:01.102656 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc49afc-b637-4c08-8318-8f9ca921569c" containerName="container-00" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.102673 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc49afc-b637-4c08-8318-8f9ca921569c" containerName="container-00" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.102919 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc49afc-b637-4c08-8318-8f9ca921569c" containerName="container-00" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.103588 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.273896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv5g8\" (UniqueName: \"kubernetes.io/projected/68f2c460-0601-44f1-bb2f-571374447b95-kube-api-access-sv5g8\") pod \"crc-debug-7mlwk\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.274331 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68f2c460-0601-44f1-bb2f-571374447b95-host\") pod \"crc-debug-7mlwk\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.342581 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc49afc-b637-4c08-8318-8f9ca921569c" path="/var/lib/kubelet/pods/7fc49afc-b637-4c08-8318-8f9ca921569c/volumes" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.375921 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68f2c460-0601-44f1-bb2f-571374447b95-host\") pod \"crc-debug-7mlwk\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.375976 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv5g8\" (UniqueName: \"kubernetes.io/projected/68f2c460-0601-44f1-bb2f-571374447b95-kube-api-access-sv5g8\") pod \"crc-debug-7mlwk\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.376065 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68f2c460-0601-44f1-bb2f-571374447b95-host\") pod \"crc-debug-7mlwk\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.398377 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv5g8\" (UniqueName: \"kubernetes.io/projected/68f2c460-0601-44f1-bb2f-571374447b95-kube-api-access-sv5g8\") pod \"crc-debug-7mlwk\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.420643 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:01 crc kubenswrapper[4775]: W1126 07:22:01.455197 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68f2c460_0601_44f1_bb2f_571374447b95.slice/crio-690a133f46e8ce9fe0b5afbe884dc097fa568bb92e5a72888099439d49cd22f7 WatchSource:0}: Error finding container 690a133f46e8ce9fe0b5afbe884dc097fa568bb92e5a72888099439d49cd22f7: Status 404 returned error can't find the container with id 690a133f46e8ce9fe0b5afbe884dc097fa568bb92e5a72888099439d49cd22f7 Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.470441 4775 scope.go:117] "RemoveContainer" containerID="da28ac7b15ad7c77dadcdc06cce1e50ca0154de6c25173000f3270c06bee7f08" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.470454 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-p24tq" Nov 26 07:22:01 crc kubenswrapper[4775]: I1126 07:22:01.472882 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-7mlwk" event={"ID":"68f2c460-0601-44f1-bb2f-571374447b95","Type":"ContainerStarted","Data":"690a133f46e8ce9fe0b5afbe884dc097fa568bb92e5a72888099439d49cd22f7"} Nov 26 07:22:02 crc kubenswrapper[4775]: I1126 07:22:02.485701 4775 generic.go:334] "Generic (PLEG): container finished" podID="68f2c460-0601-44f1-bb2f-571374447b95" containerID="5cf851f0861a638873d306fd53dc4431065e7ef1d34c13e1d2995030cdababa9" exitCode=0 Nov 26 07:22:02 crc kubenswrapper[4775]: I1126 07:22:02.485880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/crc-debug-7mlwk" event={"ID":"68f2c460-0601-44f1-bb2f-571374447b95","Type":"ContainerDied","Data":"5cf851f0861a638873d306fd53dc4431065e7ef1d34c13e1d2995030cdababa9"} Nov 26 07:22:02 crc kubenswrapper[4775]: I1126 07:22:02.526866 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48sft/crc-debug-7mlwk"] Nov 26 07:22:02 crc kubenswrapper[4775]: I1126 07:22:02.538510 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48sft/crc-debug-7mlwk"] Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.595481 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.644035 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv5g8\" (UniqueName: \"kubernetes.io/projected/68f2c460-0601-44f1-bb2f-571374447b95-kube-api-access-sv5g8\") pod \"68f2c460-0601-44f1-bb2f-571374447b95\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.644133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68f2c460-0601-44f1-bb2f-571374447b95-host\") pod \"68f2c460-0601-44f1-bb2f-571374447b95\" (UID: \"68f2c460-0601-44f1-bb2f-571374447b95\") " Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.644254 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68f2c460-0601-44f1-bb2f-571374447b95-host" (OuterVolumeSpecName: "host") pod "68f2c460-0601-44f1-bb2f-571374447b95" (UID: "68f2c460-0601-44f1-bb2f-571374447b95"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.644581 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68f2c460-0601-44f1-bb2f-571374447b95-host\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.649270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f2c460-0601-44f1-bb2f-571374447b95-kube-api-access-sv5g8" (OuterVolumeSpecName: "kube-api-access-sv5g8") pod "68f2c460-0601-44f1-bb2f-571374447b95" (UID: "68f2c460-0601-44f1-bb2f-571374447b95"). InnerVolumeSpecName "kube-api-access-sv5g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:22:03 crc kubenswrapper[4775]: I1126 07:22:03.746392 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv5g8\" (UniqueName: \"kubernetes.io/projected/68f2c460-0601-44f1-bb2f-571374447b95-kube-api-access-sv5g8\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:04 crc kubenswrapper[4775]: I1126 07:22:04.502284 4775 scope.go:117] "RemoveContainer" containerID="5cf851f0861a638873d306fd53dc4431065e7ef1d34c13e1d2995030cdababa9" Nov 26 07:22:04 crc kubenswrapper[4775]: I1126 07:22:04.502446 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/crc-debug-7mlwk" Nov 26 07:22:05 crc kubenswrapper[4775]: I1126 07:22:05.342627 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68f2c460-0601-44f1-bb2f-571374447b95" path="/var/lib/kubelet/pods/68f2c460-0601-44f1-bb2f-571374447b95/volumes" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.372955 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd8c567d4-z5vr4_1db6f2a6-c3be-4317-a6ce-802242830778/barbican-api/0.log" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.511507 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd8c567d4-z5vr4_1db6f2a6-c3be-4317-a6ce-802242830778/barbican-api-log/0.log" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.581606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-76d59b886b-lsps8_fa465504-4f8f-4c22-8f6b-3bb834f303c2/barbican-keystone-listener/0.log" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.638939 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-76d59b886b-lsps8_fa465504-4f8f-4c22-8f6b-3bb834f303c2/barbican-keystone-listener-log/0.log" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.754916 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bdbc86c75-sgb2x_2b958489-5435-42a4-bff8-577a26c717c0/barbican-worker/0.log" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.814244 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bdbc86c75-sgb2x_2b958489-5435-42a4-bff8-577a26c717c0/barbican-worker-log/0.log" Nov 26 07:22:17 crc kubenswrapper[4775]: I1126 07:22:17.941498 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz_66abe898-3799-4e9d-abc7-748423d2daa3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.079544 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/ceilometer-central-agent/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.087104 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/ceilometer-notification-agent/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.115488 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/proxy-httpd/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.222848 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/sg-core/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.303968 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b9daeb8-621b-4557-9757-6e8b9c430339/cinder-api-log/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.308130 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b9daeb8-621b-4557-9757-6e8b9c430339/cinder-api/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.501750 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3bc9a6f7-ee1c-49c1-855d-362fa796cf07/cinder-scheduler/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.559362 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3bc9a6f7-ee1c-49c1-855d-362fa796cf07/probe/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.675886 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-kp72k_7b826be9-41b3-4715-b7cb-5b8bc10e35a9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.762487 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g_c6dbfaeb-21ee-4017-aab7-6d1219d7e10f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:18 crc kubenswrapper[4775]: I1126 07:22:18.865211 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-dhkrg_e837eeaf-6cf5-44e0-b183-334f62304bc7/init/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.070562 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-dhkrg_e837eeaf-6cf5-44e0-b183-334f62304bc7/init/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.118729 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-dhkrg_e837eeaf-6cf5-44e0-b183-334f62304bc7/dnsmasq-dns/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.133901 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh_e780accc-4808-46d6-b5df-5b6492107f25/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.346457 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58818a45-1408-4416-b18e-814acd04b059/glance-log/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.353067 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58818a45-1408-4416-b18e-814acd04b059/glance-httpd/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.564839 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_46819d15-5e66-4501-b2f0-570f3b768207/glance-httpd/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.579146 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_46819d15-5e66-4501-b2f0-570f3b768207/glance-log/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.696012 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-9c9599df6-zd8gh_f578a3d5-f846-44b6-bc2b-3d98188dffff/horizon/0.log" Nov 26 07:22:19 crc kubenswrapper[4775]: I1126 07:22:19.801602 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-svnzx_0cad897c-5d2b-496c-8e45-e5061ca93a6a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:20 crc kubenswrapper[4775]: I1126 07:22:20.035428 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-9c9599df6-zd8gh_f578a3d5-f846-44b6-bc2b-3d98188dffff/horizon-log/0.log" Nov 26 07:22:20 crc kubenswrapper[4775]: I1126 07:22:20.086284 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-dndkl_22cadc11-2c9d-49f4-abf7-2e6fa3f69905/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:20 crc kubenswrapper[4775]: I1126 07:22:20.345613 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29402341-lzbf6_55d5f81a-1fab-4bfa-8ca9-115de4fef547/keystone-cron/0.log" Nov 26 07:22:20 crc kubenswrapper[4775]: I1126 07:22:20.348307 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7679ff5658-lwrpm_f06e8ee2-5844-487f-80e7-cdbba5909f74/keystone-api/0.log" Nov 26 07:22:20 crc kubenswrapper[4775]: I1126 07:22:20.478454 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2427f2bd-7928-4ff9-bc65-23bc48225184/kube-state-metrics/0.log" Nov 26 07:22:20 crc kubenswrapper[4775]: I1126 07:22:20.558590 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j_da48dd91-7390-4751-8e73-455fda656b0d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:21 crc kubenswrapper[4775]: I1126 07:22:21.045972 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54d985cc65-zsv4d_e33bc539-83a2-4077-a430-4872e8587023/neutron-api/0.log" Nov 26 07:22:21 crc kubenswrapper[4775]: I1126 07:22:21.047986 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54d985cc65-zsv4d_e33bc539-83a2-4077-a430-4872e8587023/neutron-httpd/0.log" Nov 26 07:22:21 crc kubenswrapper[4775]: I1126 07:22:21.322748 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27_51e229f8-0c28-4b0c-8307-a2dc128fa503/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:21 crc kubenswrapper[4775]: I1126 07:22:21.772026 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_76cb0fc1-4a71-4ae4-9b2e-7483efe206d7/nova-cell0-conductor-conductor/0.log" Nov 26 07:22:21 crc kubenswrapper[4775]: I1126 07:22:21.776175 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_09a96730-d166-4111-b88a-b2a51ca06b5a/nova-api-log/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.038639 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_09a96730-d166-4111-b88a-b2a51ca06b5a/nova-api-api/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.070892 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ec4083e2-96ad-420c-a66f-0cb0f274867a/nova-cell1-conductor-conductor/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.182597 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211/nova-cell1-novncproxy-novncproxy/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.307063 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ldl57_fc8364bf-1ce5-4681-99a5-edc754f142ac/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.466386 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2d57ebfc-a725-46a3-9224-f011eadfc783/nova-metadata-log/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.821119 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e8d44208-4c24-4835-89f6-a471ce50ddae/nova-scheduler-scheduler/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.900743 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c/mysql-bootstrap/0.log" Nov 26 07:22:22 crc kubenswrapper[4775]: I1126 07:22:22.994557 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c/mysql-bootstrap/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.047166 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c/galera/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.372498 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0028de9f-a7db-47c0-a552-e9f518435458/mysql-bootstrap/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.578517 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0028de9f-a7db-47c0-a552-e9f518435458/mysql-bootstrap/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.631702 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0028de9f-a7db-47c0-a552-e9f518435458/galera/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.699290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2d57ebfc-a725-46a3-9224-f011eadfc783/nova-metadata-metadata/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.792931 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3675965a-9ca7-433b-9c69-99596e74aa25/openstackclient/0.log" Nov 26 07:22:23 crc kubenswrapper[4775]: I1126 07:22:23.884706 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-k4mqj_d30382f9-28bd-4ab7-8fec-1f8932b3f804/ovn-controller/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.054616 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kd448_ff2f6960-e698-4a01-9a6d-9b1fd470e295/openstack-network-exporter/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.061495 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovsdb-server-init/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.332035 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovsdb-server/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.338479 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovs-vswitchd/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.345613 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovsdb-server-init/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.569267 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2f5d800-59f5-41fa-b45b-f2fa0981b3da/openstack-network-exporter/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.577630 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gfrt7_62bfa377-ebb1-48d4-82b6-eacd0ac68c47/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.658143 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2f5d800-59f5-41fa-b45b-f2fa0981b3da/ovn-northd/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.814667 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e61da90a-7acf-4b95-934b-69f09efb5cbc/openstack-network-exporter/0.log" Nov 26 07:22:24 crc kubenswrapper[4775]: I1126 07:22:24.923821 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e61da90a-7acf-4b95-934b-69f09efb5cbc/ovsdbserver-nb/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.021537 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dc88dc0e-14a1-479e-9682-bb5c611259a8/openstack-network-exporter/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.062441 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dc88dc0e-14a1-479e-9682-bb5c611259a8/ovsdbserver-sb/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.143977 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd45b998d-gjx5d_bc391c06-ad32-4c18-9670-481f7614e1a3/placement-api/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.349433 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd45b998d-gjx5d_bc391c06-ad32-4c18-9670-481f7614e1a3/placement-log/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.393625 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ce90c242-2ac3-4509-bf08-2f7c77b1aff0/setup-container/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.578272 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ce90c242-2ac3-4509-bf08-2f7c77b1aff0/setup-container/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.635451 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ce90c242-2ac3-4509-bf08-2f7c77b1aff0/rabbitmq/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.641353 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_837ab270-d963-4406-9a55-390bf0611e14/setup-container/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.832292 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_837ab270-d963-4406-9a55-390bf0611e14/rabbitmq/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.835389 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_837ab270-d963-4406-9a55-390bf0611e14/setup-container/0.log" Nov 26 07:22:25 crc kubenswrapper[4775]: I1126 07:22:25.854598 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl_efef5423-9419-4ff8-b603-96ffaa7f4dcd/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.065457 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57_561d8a6c-3d90-4220-b6ac-386d81f5016f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.103447 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-gm5w2_f949db3e-70cb-458b-96d9-9c9f0f1118d8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.309053 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lljnf_6233c3cd-e8bc-4dd1-b039-c950e119ce4e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.402659 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ghkc9_7e9b9292-e385-4016-8355-2b95cf0e85af/ssh-known-hosts-edpm-deployment/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.593909 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6c8b9f5bf9-nsz2f_546c7a7f-b54b-415e-81c6-0fb3fa734e05/proxy-server/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.638191 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6c8b9f5bf9-nsz2f_546c7a7f-b54b-415e-81c6-0fb3fa734e05/proxy-httpd/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.818123 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-d8msj_fd3c60bd-b664-43b1-b669-3ef640959006/swift-ring-rebalance/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.878945 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-auditor/0.log" Nov 26 07:22:26 crc kubenswrapper[4775]: I1126 07:22:26.881621 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-reaper/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.028934 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-replicator/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.083969 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-server/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.123224 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-replicator/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.134155 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-auditor/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.230023 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-server/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.259350 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-updater/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.363129 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-auditor/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.372933 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-expirer/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.452748 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-replicator/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.507115 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-server/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.579948 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/rsync/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.581749 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-updater/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.652578 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/swift-recon-cron/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.892756 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2ssps_1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:27 crc kubenswrapper[4775]: I1126 07:22:27.953357 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5/tempest-tests-tempest-tests-runner/0.log" Nov 26 07:22:28 crc kubenswrapper[4775]: I1126 07:22:28.105700 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7075bbd0-afa7-4430-b2b9-ee07eac73b2a/test-operator-logs-container/0.log" Nov 26 07:22:28 crc kubenswrapper[4775]: I1126 07:22:28.197283 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-z74zq_3d9ee8c3-9727-4454-8df5-5ecce45daf45/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:22:36 crc kubenswrapper[4775]: I1126 07:22:36.525224 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bf022b43-4760-4e21-8f5b-2558a69ede14/memcached/0.log" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.405019 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wjhxh"] Nov 26 07:22:46 crc kubenswrapper[4775]: E1126 07:22:46.406058 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f2c460-0601-44f1-bb2f-571374447b95" containerName="container-00" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.406074 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f2c460-0601-44f1-bb2f-571374447b95" containerName="container-00" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.406282 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f2c460-0601-44f1-bb2f-571374447b95" containerName="container-00" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.407985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.427838 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wjhxh"] Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.459582 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-utilities\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.459981 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zjfp\" (UniqueName: \"kubernetes.io/projected/e77eacbd-64fb-451c-b51d-19e19ee48abb-kube-api-access-2zjfp\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.460029 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-catalog-content\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.561523 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-utilities\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.561724 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zjfp\" (UniqueName: \"kubernetes.io/projected/e77eacbd-64fb-451c-b51d-19e19ee48abb-kube-api-access-2zjfp\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.561753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-catalog-content\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.562071 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-utilities\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.562333 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-catalog-content\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.584063 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zjfp\" (UniqueName: \"kubernetes.io/projected/e77eacbd-64fb-451c-b51d-19e19ee48abb-kube-api-access-2zjfp\") pod \"redhat-marketplace-wjhxh\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:46 crc kubenswrapper[4775]: I1126 07:22:46.737439 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:47 crc kubenswrapper[4775]: I1126 07:22:47.297256 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wjhxh"] Nov 26 07:22:47 crc kubenswrapper[4775]: W1126 07:22:47.297854 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode77eacbd_64fb_451c_b51d_19e19ee48abb.slice/crio-12a326eec80f1196a8c5d1037b46cb3750a148cb8d02cb68d9072e62cf20b0a0 WatchSource:0}: Error finding container 12a326eec80f1196a8c5d1037b46cb3750a148cb8d02cb68d9072e62cf20b0a0: Status 404 returned error can't find the container with id 12a326eec80f1196a8c5d1037b46cb3750a148cb8d02cb68d9072e62cf20b0a0 Nov 26 07:22:47 crc kubenswrapper[4775]: I1126 07:22:47.877517 4775 generic.go:334] "Generic (PLEG): container finished" podID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerID="4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89" exitCode=0 Nov 26 07:22:47 crc kubenswrapper[4775]: I1126 07:22:47.877583 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wjhxh" event={"ID":"e77eacbd-64fb-451c-b51d-19e19ee48abb","Type":"ContainerDied","Data":"4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89"} Nov 26 07:22:47 crc kubenswrapper[4775]: I1126 07:22:47.877829 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wjhxh" event={"ID":"e77eacbd-64fb-451c-b51d-19e19ee48abb","Type":"ContainerStarted","Data":"12a326eec80f1196a8c5d1037b46cb3750a148cb8d02cb68d9072e62cf20b0a0"} Nov 26 07:22:49 crc kubenswrapper[4775]: I1126 07:22:49.897595 4775 generic.go:334] "Generic (PLEG): container finished" podID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerID="0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55" exitCode=0 Nov 26 07:22:49 crc kubenswrapper[4775]: I1126 07:22:49.897764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wjhxh" event={"ID":"e77eacbd-64fb-451c-b51d-19e19ee48abb","Type":"ContainerDied","Data":"0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55"} Nov 26 07:22:50 crc kubenswrapper[4775]: I1126 07:22:50.909429 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wjhxh" event={"ID":"e77eacbd-64fb-451c-b51d-19e19ee48abb","Type":"ContainerStarted","Data":"0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda"} Nov 26 07:22:50 crc kubenswrapper[4775]: I1126 07:22:50.935344 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wjhxh" podStartSLOduration=2.522060431 podStartE2EDuration="4.935328227s" podCreationTimestamp="2025-11-26 07:22:46 +0000 UTC" firstStartedPulling="2025-11-26 07:22:47.880132005 +0000 UTC m=+3451.241435957" lastFinishedPulling="2025-11-26 07:22:50.293399801 +0000 UTC m=+3453.654703753" observedRunningTime="2025-11-26 07:22:50.929708587 +0000 UTC m=+3454.291012539" watchObservedRunningTime="2025-11-26 07:22:50.935328227 +0000 UTC m=+3454.296632179" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.292100 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/util/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.434691 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/util/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.541444 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/pull/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.550961 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/pull/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.761776 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/util/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.780096 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/pull/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.788652 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/extract/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.932223 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-27cqk_4f5effa7-3458-4978-8382-0f5b4a17105a/kube-rbac-proxy/0.log" Nov 26 07:22:51 crc kubenswrapper[4775]: I1126 07:22:51.989732 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-5whjf_153b27e4-d948-4ff6-9d6a-bfbcebd17cad/kube-rbac-proxy/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.023336 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-27cqk_4f5effa7-3458-4978-8382-0f5b4a17105a/manager/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.181885 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-5whjf_153b27e4-d948-4ff6-9d6a-bfbcebd17cad/manager/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.242531 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dghfx_1a1d393e-2d6c-425c-9e6f-068d4c8ee090/manager/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.256963 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dghfx_1a1d393e-2d6c-425c-9e6f-068d4c8ee090/kube-rbac-proxy/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.382476 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-sdz7q_cc351f29-006f-4ee4-80cb-5860f4f824ff/kube-rbac-proxy/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.532470 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-sdz7q_cc351f29-006f-4ee4-80cb-5860f4f824ff/manager/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.585753 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-d7m5d_faa5a2e8-ca2b-42e1-ac9a-3580111707ac/kube-rbac-proxy/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.621025 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-d7m5d_faa5a2e8-ca2b-42e1-ac9a-3580111707ac/manager/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.704667 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-jmwr9_ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7/kube-rbac-proxy/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.781449 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-jmwr9_ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7/manager/0.log" Nov 26 07:22:52 crc kubenswrapper[4775]: I1126 07:22:52.897240 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-xrzlh_8483a9f8-d70d-497d-ae02-c6b426a49306/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.068913 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-gdk4v_2158b3e8-3c08-4e7c-9b28-cd32534135a1/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.073693 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-xrzlh_8483a9f8-d70d-497d-ae02-c6b426a49306/manager/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.140649 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-gdk4v_2158b3e8-3c08-4e7c-9b28-cd32534135a1/manager/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.253798 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4fxcp_997d00b8-4cc3-45e9-8af2-2d573eef844a/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.322301 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4fxcp_997d00b8-4cc3-45e9-8af2-2d573eef844a/manager/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.451526 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-5hvj7_021e1511-a060-4d30-af62-872ba84b84c3/manager/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.454126 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-5hvj7_021e1511-a060-4d30-af62-872ba84b84c3/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.531655 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-6jrgg_85c90cee-6304-4ab9-957c-0bd91411403f/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.626559 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-6jrgg_85c90cee-6304-4ab9-957c-0bd91411403f/manager/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.689203 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-q97hn_7bead1ee-7a42-45c9-9205-420ae85002f7/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.749357 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-q97hn_7bead1ee-7a42-45c9-9205-420ae85002f7/manager/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.891640 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ftsgz_d8ad3c0b-931e-444d-9fcf-23dfbe77bedc/kube-rbac-proxy/0.log" Nov 26 07:22:53 crc kubenswrapper[4775]: I1126 07:22:53.952537 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ftsgz_d8ad3c0b-931e-444d-9fcf-23dfbe77bedc/manager/0.log" Nov 26 07:22:54 crc kubenswrapper[4775]: I1126 07:22:54.061493 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-plpx4_d22fe1a8-9183-4ca4-ae47-60d3e77f2339/kube-rbac-proxy/0.log" Nov 26 07:22:54 crc kubenswrapper[4775]: I1126 07:22:54.084141 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-plpx4_d22fe1a8-9183-4ca4-ae47-60d3e77f2339/manager/0.log" Nov 26 07:22:54 crc kubenswrapper[4775]: I1126 07:22:54.190431 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv_7ac95117-0ba0-4c51-afbf-320322b8ebfc/kube-rbac-proxy/0.log" Nov 26 07:22:54 crc kubenswrapper[4775]: I1126 07:22:54.305418 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv_7ac95117-0ba0-4c51-afbf-320322b8ebfc/manager/0.log" Nov 26 07:22:54 crc kubenswrapper[4775]: I1126 07:22:54.738065 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6df95dcb7f-5wnrd_2b412dfe-90b0-4018-81ca-32d9644fbc99/operator/0.log" Nov 26 07:22:54 crc kubenswrapper[4775]: I1126 07:22:54.898452 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rpzlv_cd5862a0-b9ab-4473-bbe4-34253c8cbb79/registry-server/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.041695 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-ksct6_5e867f5f-7847-4469-95cb-605042211f56/kube-rbac-proxy/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.360754 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-ffw7r_f9b85bd6-21c3-457e-81c0-b87c56be9f3e/kube-rbac-proxy/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.402901 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-ksct6_5e867f5f-7847-4469-95cb-605042211f56/manager/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.467030 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-ffw7r_f9b85bd6-21c3-457e-81c0-b87c56be9f3e/manager/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.607268 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-r7dqc_2e76307b-7079-41d7-a66e-35f03cc1a1bd/operator/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.671300 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2s825_e72bf30c-6c74-46a6-b215-ac5b1f33c144/kube-rbac-proxy/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.741580 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-646778448f-7d5mc_cf2135a7-5206-498c-9391-379fae5f4bfa/manager/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.845388 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2s825_e72bf30c-6c74-46a6-b215-ac5b1f33c144/manager/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.890482 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fkb27_04b644f3-2f69-4cb5-8ade-04f4be267255/kube-rbac-proxy/0.log" Nov 26 07:22:55 crc kubenswrapper[4775]: I1126 07:22:55.973399 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fkb27_04b644f3-2f69-4cb5-8ade-04f4be267255/manager/0.log" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.042635 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-8sjts_523ab37c-4bb4-4d00-8134-ad63c2833907/kube-rbac-proxy/0.log" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.047878 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-8sjts_523ab37c-4bb4-4d00-8134-ad63c2833907/manager/0.log" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.155753 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-65qdq_0fa640fb-a493-4f84-bfcf-a3824ddc061f/kube-rbac-proxy/0.log" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.233983 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-65qdq_0fa640fb-a493-4f84-bfcf-a3824ddc061f/manager/0.log" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.738400 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.738449 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:56 crc kubenswrapper[4775]: I1126 07:22:56.806343 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:57 crc kubenswrapper[4775]: I1126 07:22:57.005077 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:57 crc kubenswrapper[4775]: I1126 07:22:57.055915 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wjhxh"] Nov 26 07:22:58 crc kubenswrapper[4775]: I1126 07:22:58.969929 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wjhxh" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="registry-server" containerID="cri-o://0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda" gracePeriod=2 Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.416611 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.534056 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-catalog-content\") pod \"e77eacbd-64fb-451c-b51d-19e19ee48abb\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.534145 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-utilities\") pod \"e77eacbd-64fb-451c-b51d-19e19ee48abb\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.534182 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zjfp\" (UniqueName: \"kubernetes.io/projected/e77eacbd-64fb-451c-b51d-19e19ee48abb-kube-api-access-2zjfp\") pod \"e77eacbd-64fb-451c-b51d-19e19ee48abb\" (UID: \"e77eacbd-64fb-451c-b51d-19e19ee48abb\") " Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.535665 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-utilities" (OuterVolumeSpecName: "utilities") pod "e77eacbd-64fb-451c-b51d-19e19ee48abb" (UID: "e77eacbd-64fb-451c-b51d-19e19ee48abb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.535997 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.547979 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77eacbd-64fb-451c-b51d-19e19ee48abb-kube-api-access-2zjfp" (OuterVolumeSpecName: "kube-api-access-2zjfp") pod "e77eacbd-64fb-451c-b51d-19e19ee48abb" (UID: "e77eacbd-64fb-451c-b51d-19e19ee48abb"). InnerVolumeSpecName "kube-api-access-2zjfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.558308 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e77eacbd-64fb-451c-b51d-19e19ee48abb" (UID: "e77eacbd-64fb-451c-b51d-19e19ee48abb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.637851 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e77eacbd-64fb-451c-b51d-19e19ee48abb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.637890 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zjfp\" (UniqueName: \"kubernetes.io/projected/e77eacbd-64fb-451c-b51d-19e19ee48abb-kube-api-access-2zjfp\") on node \"crc\" DevicePath \"\"" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.979972 4775 generic.go:334] "Generic (PLEG): container finished" podID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerID="0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda" exitCode=0 Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.980030 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wjhxh" Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.980024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wjhxh" event={"ID":"e77eacbd-64fb-451c-b51d-19e19ee48abb","Type":"ContainerDied","Data":"0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda"} Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.980103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wjhxh" event={"ID":"e77eacbd-64fb-451c-b51d-19e19ee48abb","Type":"ContainerDied","Data":"12a326eec80f1196a8c5d1037b46cb3750a148cb8d02cb68d9072e62cf20b0a0"} Nov 26 07:22:59 crc kubenswrapper[4775]: I1126 07:22:59.980124 4775 scope.go:117] "RemoveContainer" containerID="0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.001121 4775 scope.go:117] "RemoveContainer" containerID="0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.017512 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wjhxh"] Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.025329 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wjhxh"] Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.037627 4775 scope.go:117] "RemoveContainer" containerID="4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.069494 4775 scope.go:117] "RemoveContainer" containerID="0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda" Nov 26 07:23:00 crc kubenswrapper[4775]: E1126 07:23:00.069961 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda\": container with ID starting with 0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda not found: ID does not exist" containerID="0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.070025 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda"} err="failed to get container status \"0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda\": rpc error: code = NotFound desc = could not find container \"0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda\": container with ID starting with 0b189e6c0133f280cb91d6962df8df24f7dddeb431f307900031b71b17bd3bda not found: ID does not exist" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.070047 4775 scope.go:117] "RemoveContainer" containerID="0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55" Nov 26 07:23:00 crc kubenswrapper[4775]: E1126 07:23:00.070395 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55\": container with ID starting with 0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55 not found: ID does not exist" containerID="0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.070417 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55"} err="failed to get container status \"0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55\": rpc error: code = NotFound desc = could not find container \"0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55\": container with ID starting with 0fb289590017ab93611e7608032d65af60c2d4977dc71e12a1d8fed6dab06b55 not found: ID does not exist" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.070434 4775 scope.go:117] "RemoveContainer" containerID="4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89" Nov 26 07:23:00 crc kubenswrapper[4775]: E1126 07:23:00.070771 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89\": container with ID starting with 4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89 not found: ID does not exist" containerID="4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.070822 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89"} err="failed to get container status \"4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89\": rpc error: code = NotFound desc = could not find container \"4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89\": container with ID starting with 4c720a58b4e1fd4d63855d57089fc2742b631c583290ff89b60ab00209fb6a89 not found: ID does not exist" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.260387 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4vx4j"] Nov 26 07:23:00 crc kubenswrapper[4775]: E1126 07:23:00.260860 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="registry-server" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.260878 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="registry-server" Nov 26 07:23:00 crc kubenswrapper[4775]: E1126 07:23:00.260908 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="extract-utilities" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.260917 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="extract-utilities" Nov 26 07:23:00 crc kubenswrapper[4775]: E1126 07:23:00.260945 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="extract-content" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.260955 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="extract-content" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.261205 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" containerName="registry-server" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.263361 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.276123 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4vx4j"] Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.350078 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-catalog-content\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.350124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zlj6\" (UniqueName: \"kubernetes.io/projected/572e0013-9c3b-4690-90b2-06eefae2e3cd-kube-api-access-7zlj6\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.350332 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-utilities\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.452585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-catalog-content\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.452635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zlj6\" (UniqueName: \"kubernetes.io/projected/572e0013-9c3b-4690-90b2-06eefae2e3cd-kube-api-access-7zlj6\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.452699 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-utilities\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.453166 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-catalog-content\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.453211 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-utilities\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.472157 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zlj6\" (UniqueName: \"kubernetes.io/projected/572e0013-9c3b-4690-90b2-06eefae2e3cd-kube-api-access-7zlj6\") pod \"community-operators-4vx4j\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.632015 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.872913 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-65b8x"] Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.876784 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.901935 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-65b8x"] Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.961246 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-utilities\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.961401 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-catalog-content\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:00 crc kubenswrapper[4775]: I1126 07:23:00.961439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gll88\" (UniqueName: \"kubernetes.io/projected/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-kube-api-access-gll88\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.063054 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-catalog-content\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.063113 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gll88\" (UniqueName: \"kubernetes.io/projected/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-kube-api-access-gll88\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.063188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-utilities\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.063671 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-utilities\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.063891 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-catalog-content\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.092147 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gll88\" (UniqueName: \"kubernetes.io/projected/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-kube-api-access-gll88\") pod \"redhat-operators-65b8x\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.204996 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4vx4j"] Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.243550 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.382557 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e77eacbd-64fb-451c-b51d-19e19ee48abb" path="/var/lib/kubelet/pods/e77eacbd-64fb-451c-b51d-19e19ee48abb/volumes" Nov 26 07:23:01 crc kubenswrapper[4775]: I1126 07:23:01.735294 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-65b8x"] Nov 26 07:23:01 crc kubenswrapper[4775]: W1126 07:23:01.746904 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b8d6a2_d672_4c1c_9ff6_aba6a2f199c0.slice/crio-618c2beaa9b2e313003c1aa2fbf1d914bf8e1cf625822de84b16fb9c24589571 WatchSource:0}: Error finding container 618c2beaa9b2e313003c1aa2fbf1d914bf8e1cf625822de84b16fb9c24589571: Status 404 returned error can't find the container with id 618c2beaa9b2e313003c1aa2fbf1d914bf8e1cf625822de84b16fb9c24589571 Nov 26 07:23:02 crc kubenswrapper[4775]: I1126 07:23:02.006385 4775 generic.go:334] "Generic (PLEG): container finished" podID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerID="65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37" exitCode=0 Nov 26 07:23:02 crc kubenswrapper[4775]: I1126 07:23:02.006549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerDied","Data":"65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37"} Nov 26 07:23:02 crc kubenswrapper[4775]: I1126 07:23:02.006776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerStarted","Data":"258e68bb3d1eec4f2a878e9e13db13f041561009d69f9b6451b656a1e15893d8"} Nov 26 07:23:02 crc kubenswrapper[4775]: I1126 07:23:02.009033 4775 generic.go:334] "Generic (PLEG): container finished" podID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerID="7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9" exitCode=0 Nov 26 07:23:02 crc kubenswrapper[4775]: I1126 07:23:02.009100 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerDied","Data":"7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9"} Nov 26 07:23:02 crc kubenswrapper[4775]: I1126 07:23:02.009129 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerStarted","Data":"618c2beaa9b2e313003c1aa2fbf1d914bf8e1cf625822de84b16fb9c24589571"} Nov 26 07:23:03 crc kubenswrapper[4775]: I1126 07:23:03.039656 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerStarted","Data":"4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87"} Nov 26 07:23:03 crc kubenswrapper[4775]: I1126 07:23:03.043939 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerStarted","Data":"5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40"} Nov 26 07:23:04 crc kubenswrapper[4775]: I1126 07:23:04.057645 4775 generic.go:334] "Generic (PLEG): container finished" podID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerID="4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87" exitCode=0 Nov 26 07:23:04 crc kubenswrapper[4775]: I1126 07:23:04.057729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerDied","Data":"4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87"} Nov 26 07:23:04 crc kubenswrapper[4775]: I1126 07:23:04.061016 4775 generic.go:334] "Generic (PLEG): container finished" podID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerID="5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40" exitCode=0 Nov 26 07:23:04 crc kubenswrapper[4775]: I1126 07:23:04.061066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerDied","Data":"5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40"} Nov 26 07:23:06 crc kubenswrapper[4775]: I1126 07:23:06.080420 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerStarted","Data":"d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426"} Nov 26 07:23:06 crc kubenswrapper[4775]: I1126 07:23:06.082864 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerStarted","Data":"1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a"} Nov 26 07:23:06 crc kubenswrapper[4775]: I1126 07:23:06.098273 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4vx4j" podStartSLOduration=2.96635128 podStartE2EDuration="6.098259529s" podCreationTimestamp="2025-11-26 07:23:00 +0000 UTC" firstStartedPulling="2025-11-26 07:23:02.008581848 +0000 UTC m=+3465.369885790" lastFinishedPulling="2025-11-26 07:23:05.140490087 +0000 UTC m=+3468.501794039" observedRunningTime="2025-11-26 07:23:06.097059817 +0000 UTC m=+3469.458363779" watchObservedRunningTime="2025-11-26 07:23:06.098259529 +0000 UTC m=+3469.459563481" Nov 26 07:23:06 crc kubenswrapper[4775]: I1126 07:23:06.123817 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-65b8x" podStartSLOduration=3.104712462 podStartE2EDuration="6.123791581s" podCreationTimestamp="2025-11-26 07:23:00 +0000 UTC" firstStartedPulling="2025-11-26 07:23:02.010269363 +0000 UTC m=+3465.371573315" lastFinishedPulling="2025-11-26 07:23:05.029348482 +0000 UTC m=+3468.390652434" observedRunningTime="2025-11-26 07:23:06.118252313 +0000 UTC m=+3469.479556285" watchObservedRunningTime="2025-11-26 07:23:06.123791581 +0000 UTC m=+3469.485095533" Nov 26 07:23:10 crc kubenswrapper[4775]: I1126 07:23:10.632726 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:10 crc kubenswrapper[4775]: I1126 07:23:10.634279 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:10 crc kubenswrapper[4775]: I1126 07:23:10.680047 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:11 crc kubenswrapper[4775]: I1126 07:23:11.244523 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:11 crc kubenswrapper[4775]: I1126 07:23:11.244594 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:11 crc kubenswrapper[4775]: I1126 07:23:11.295942 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:11 crc kubenswrapper[4775]: I1126 07:23:11.713143 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:11 crc kubenswrapper[4775]: I1126 07:23:11.713926 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:11 crc kubenswrapper[4775]: I1126 07:23:11.912223 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4vx4j"] Nov 26 07:23:13 crc kubenswrapper[4775]: I1126 07:23:13.682275 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4vx4j" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="registry-server" containerID="cri-o://d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426" gracePeriod=2 Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.119412 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-65b8x"] Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.119979 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-65b8x" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="registry-server" containerID="cri-o://1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a" gracePeriod=2 Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.142532 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.316905 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zlj6\" (UniqueName: \"kubernetes.io/projected/572e0013-9c3b-4690-90b2-06eefae2e3cd-kube-api-access-7zlj6\") pod \"572e0013-9c3b-4690-90b2-06eefae2e3cd\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.317023 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-utilities\") pod \"572e0013-9c3b-4690-90b2-06eefae2e3cd\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.317188 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-catalog-content\") pod \"572e0013-9c3b-4690-90b2-06eefae2e3cd\" (UID: \"572e0013-9c3b-4690-90b2-06eefae2e3cd\") " Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.317889 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-utilities" (OuterVolumeSpecName: "utilities") pod "572e0013-9c3b-4690-90b2-06eefae2e3cd" (UID: "572e0013-9c3b-4690-90b2-06eefae2e3cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.320917 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.351486 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572e0013-9c3b-4690-90b2-06eefae2e3cd-kube-api-access-7zlj6" (OuterVolumeSpecName: "kube-api-access-7zlj6") pod "572e0013-9c3b-4690-90b2-06eefae2e3cd" (UID: "572e0013-9c3b-4690-90b2-06eefae2e3cd"). InnerVolumeSpecName "kube-api-access-7zlj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.369110 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "572e0013-9c3b-4690-90b2-06eefae2e3cd" (UID: "572e0013-9c3b-4690-90b2-06eefae2e3cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.423394 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572e0013-9c3b-4690-90b2-06eefae2e3cd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.423434 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zlj6\" (UniqueName: \"kubernetes.io/projected/572e0013-9c3b-4690-90b2-06eefae2e3cd-kube-api-access-7zlj6\") on node \"crc\" DevicePath \"\"" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.485110 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.629058 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gll88\" (UniqueName: \"kubernetes.io/projected/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-kube-api-access-gll88\") pod \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.629178 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-utilities\") pod \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.629435 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-catalog-content\") pod \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\" (UID: \"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0\") " Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.629871 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-utilities" (OuterVolumeSpecName: "utilities") pod "63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" (UID: "63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.632419 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-kube-api-access-gll88" (OuterVolumeSpecName: "kube-api-access-gll88") pod "63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" (UID: "63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0"). InnerVolumeSpecName "kube-api-access-gll88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.692407 4775 generic.go:334] "Generic (PLEG): container finished" podID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerID="d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426" exitCode=0 Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.692477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerDied","Data":"d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426"} Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.692509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4vx4j" event={"ID":"572e0013-9c3b-4690-90b2-06eefae2e3cd","Type":"ContainerDied","Data":"258e68bb3d1eec4f2a878e9e13db13f041561009d69f9b6451b656a1e15893d8"} Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.692528 4775 scope.go:117] "RemoveContainer" containerID="d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.692686 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4vx4j" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.704063 4775 generic.go:334] "Generic (PLEG): container finished" podID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerID="1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a" exitCode=0 Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.704113 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerDied","Data":"1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a"} Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.704141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-65b8x" event={"ID":"63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0","Type":"ContainerDied","Data":"618c2beaa9b2e313003c1aa2fbf1d914bf8e1cf625822de84b16fb9c24589571"} Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.705300 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-65b8x" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.719769 4775 scope.go:117] "RemoveContainer" containerID="4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.730601 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" (UID: "63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.732314 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.732347 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gll88\" (UniqueName: \"kubernetes.io/projected/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-kube-api-access-gll88\") on node \"crc\" DevicePath \"\"" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.732361 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.738767 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4vx4j"] Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.746734 4775 scope.go:117] "RemoveContainer" containerID="65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.748644 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4vx4j"] Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.765908 4775 scope.go:117] "RemoveContainer" containerID="d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426" Nov 26 07:23:14 crc kubenswrapper[4775]: E1126 07:23:14.766310 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426\": container with ID starting with d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426 not found: ID does not exist" containerID="d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.766350 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426"} err="failed to get container status \"d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426\": rpc error: code = NotFound desc = could not find container \"d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426\": container with ID starting with d491e3cf88b1f3710de9a606dfb99efc1334bd7d81044c67852feb2de1631426 not found: ID does not exist" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.766372 4775 scope.go:117] "RemoveContainer" containerID="4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87" Nov 26 07:23:14 crc kubenswrapper[4775]: E1126 07:23:14.766947 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87\": container with ID starting with 4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87 not found: ID does not exist" containerID="4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.767008 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87"} err="failed to get container status \"4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87\": rpc error: code = NotFound desc = could not find container \"4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87\": container with ID starting with 4eb4ad6f456e6f20f5b938cd5d977b73aa12c0c2f056bf98038a3b636de47f87 not found: ID does not exist" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.767062 4775 scope.go:117] "RemoveContainer" containerID="65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37" Nov 26 07:23:14 crc kubenswrapper[4775]: E1126 07:23:14.767902 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37\": container with ID starting with 65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37 not found: ID does not exist" containerID="65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.767935 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37"} err="failed to get container status \"65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37\": rpc error: code = NotFound desc = could not find container \"65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37\": container with ID starting with 65ad64d2ee8181352469799a6f49c41b9f733be1c01a760582720790f4af7d37 not found: ID does not exist" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.767954 4775 scope.go:117] "RemoveContainer" containerID="1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.786114 4775 scope.go:117] "RemoveContainer" containerID="5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.860262 4775 scope.go:117] "RemoveContainer" containerID="7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.916050 4775 scope.go:117] "RemoveContainer" containerID="1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a" Nov 26 07:23:14 crc kubenswrapper[4775]: E1126 07:23:14.920456 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a\": container with ID starting with 1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a not found: ID does not exist" containerID="1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.920507 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a"} err="failed to get container status \"1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a\": rpc error: code = NotFound desc = could not find container \"1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a\": container with ID starting with 1a0a28e9afe9a8d1bcede57865f2908afcbb6847013763f4b7cc2965b245942a not found: ID does not exist" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.920538 4775 scope.go:117] "RemoveContainer" containerID="5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40" Nov 26 07:23:14 crc kubenswrapper[4775]: E1126 07:23:14.920906 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40\": container with ID starting with 5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40 not found: ID does not exist" containerID="5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.920935 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40"} err="failed to get container status \"5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40\": rpc error: code = NotFound desc = could not find container \"5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40\": container with ID starting with 5364313e7ddfda583569386099be8b124ef78bb702c633a043756eb463386e40 not found: ID does not exist" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.920950 4775 scope.go:117] "RemoveContainer" containerID="7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9" Nov 26 07:23:14 crc kubenswrapper[4775]: E1126 07:23:14.921162 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9\": container with ID starting with 7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9 not found: ID does not exist" containerID="7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9" Nov 26 07:23:14 crc kubenswrapper[4775]: I1126 07:23:14.921188 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9"} err="failed to get container status \"7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9\": rpc error: code = NotFound desc = could not find container \"7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9\": container with ID starting with 7a5b92f9c2f5505bc4a3228fdbf1a328c20cd4f6d4b62dcadcd220a920c2cfb9 not found: ID does not exist" Nov 26 07:23:15 crc kubenswrapper[4775]: I1126 07:23:15.041814 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-65b8x"] Nov 26 07:23:15 crc kubenswrapper[4775]: I1126 07:23:15.053743 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-65b8x"] Nov 26 07:23:15 crc kubenswrapper[4775]: I1126 07:23:15.338244 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" path="/var/lib/kubelet/pods/572e0013-9c3b-4690-90b2-06eefae2e3cd/volumes" Nov 26 07:23:15 crc kubenswrapper[4775]: I1126 07:23:15.338915 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" path="/var/lib/kubelet/pods/63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0/volumes" Nov 26 07:23:16 crc kubenswrapper[4775]: I1126 07:23:16.737027 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-x5x92_cb8fd937-d759-4e21-98af-5b833a8b3f52/control-plane-machine-set-operator/0.log" Nov 26 07:23:16 crc kubenswrapper[4775]: I1126 07:23:16.917080 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p4qsm_74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb/kube-rbac-proxy/0.log" Nov 26 07:23:16 crc kubenswrapper[4775]: I1126 07:23:16.920733 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p4qsm_74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb/machine-api-operator/0.log" Nov 26 07:23:28 crc kubenswrapper[4775]: I1126 07:23:28.220722 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-f9chx_4fe2d6c0-b093-452f-8dfa-0db642716f10/cert-manager-controller/0.log" Nov 26 07:23:28 crc kubenswrapper[4775]: I1126 07:23:28.456594 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-99ngt_1d540303-134e-4ca8-9597-9eb8dda29ef4/cert-manager-cainjector/0.log" Nov 26 07:23:28 crc kubenswrapper[4775]: I1126 07:23:28.537443 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-mn5gf_8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd/cert-manager-webhook/0.log" Nov 26 07:23:40 crc kubenswrapper[4775]: I1126 07:23:40.676013 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-k5dwr_d1e0c343-feef-4493-bec5-d725f665b9d6/nmstate-console-plugin/0.log" Nov 26 07:23:40 crc kubenswrapper[4775]: I1126 07:23:40.875049 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2tzgw_98657427-74ec-48e6-89d9-ef27646c475a/nmstate-handler/0.log" Nov 26 07:23:40 crc kubenswrapper[4775]: I1126 07:23:40.957153 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-2xn6f_910b3c41-e6e8-4c07-945d-d9c580ccce6c/kube-rbac-proxy/0.log" Nov 26 07:23:41 crc kubenswrapper[4775]: I1126 07:23:41.040012 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-2xn6f_910b3c41-e6e8-4c07-945d-d9c580ccce6c/nmstate-metrics/0.log" Nov 26 07:23:41 crc kubenswrapper[4775]: I1126 07:23:41.143684 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-jhrbn_5b010a92-1985-467a-9d70-35d65f48f13d/nmstate-operator/0.log" Nov 26 07:23:41 crc kubenswrapper[4775]: I1126 07:23:41.226185 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-c49w5_2f477e8a-59b8-460b-b84d-6b1a4fd50827/nmstate-webhook/0.log" Nov 26 07:23:51 crc kubenswrapper[4775]: I1126 07:23:51.419976 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:23:51 crc kubenswrapper[4775]: I1126 07:23:51.420503 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.094095 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-lnhds_2f366115-1210-4d05-9f99-b00eede5cee4/kube-rbac-proxy/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.287831 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-lnhds_2f366115-1210-4d05-9f99-b00eede5cee4/controller/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.317453 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.463454 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.546615 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.558320 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.576789 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.763133 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.772036 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.779341 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.816652 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:23:55 crc kubenswrapper[4775]: I1126 07:23:55.986668 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.003959 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/controller/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.011858 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.031512 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.152176 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/frr-metrics/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.205904 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/kube-rbac-proxy-frr/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.244030 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/kube-rbac-proxy/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.376643 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/reloader/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.502982 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-r6bx2_dccc8029-10ed-427e-a26c-74427323350b/frr-k8s-webhook-server/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.687488 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79f98797bf-xz98r_8e93463e-dc10-48b2-b359-494a53052995/manager/0.log" Nov 26 07:23:56 crc kubenswrapper[4775]: I1126 07:23:56.878134 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-766d5f58fc-d4gll_e0758bd1-82f1-4a2f-a009-75975267444c/webhook-server/0.log" Nov 26 07:23:57 crc kubenswrapper[4775]: I1126 07:23:57.051751 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42hxm_953ed086-d444-4c86-9217-134ca8126184/kube-rbac-proxy/0.log" Nov 26 07:23:57 crc kubenswrapper[4775]: I1126 07:23:57.640574 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/frr/0.log" Nov 26 07:23:57 crc kubenswrapper[4775]: I1126 07:23:57.640982 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42hxm_953ed086-d444-4c86-9217-134ca8126184/speaker/0.log" Nov 26 07:24:10 crc kubenswrapper[4775]: I1126 07:24:10.716517 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/util/0.log" Nov 26 07:24:10 crc kubenswrapper[4775]: I1126 07:24:10.884587 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/util/0.log" Nov 26 07:24:10 crc kubenswrapper[4775]: I1126 07:24:10.913080 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/pull/0.log" Nov 26 07:24:10 crc kubenswrapper[4775]: I1126 07:24:10.922270 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/pull/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.089244 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/pull/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.095067 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/util/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.095246 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/extract/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.254226 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-utilities/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.425210 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-content/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.428553 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-content/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.431151 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-utilities/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.632655 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-utilities/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.635068 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-content/0.log" Nov 26 07:24:11 crc kubenswrapper[4775]: I1126 07:24:11.894696 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-utilities/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.027312 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-utilities/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.112504 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-content/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.146322 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-content/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.206812 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/registry-server/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.336139 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-utilities/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.353488 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-content/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.572628 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/util/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.890932 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/registry-server/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.918158 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/pull/0.log" Nov 26 07:24:12 crc kubenswrapper[4775]: I1126 07:24:12.941263 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/pull/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.057834 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/util/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.122897 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/util/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.145892 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/pull/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.181864 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/extract/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.322330 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sh8wn_a4c5e343-455f-43ad-9936-8b64224702bf/marketplace-operator/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.408504 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-utilities/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.586760 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-utilities/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.593626 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-content/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.633039 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-content/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.761553 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-utilities/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.773546 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-content/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.948453 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/registry-server/0.log" Nov 26 07:24:13 crc kubenswrapper[4775]: I1126 07:24:13.981582 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-utilities/0.log" Nov 26 07:24:14 crc kubenswrapper[4775]: I1126 07:24:14.133150 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-content/0.log" Nov 26 07:24:14 crc kubenswrapper[4775]: I1126 07:24:14.164608 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-utilities/0.log" Nov 26 07:24:14 crc kubenswrapper[4775]: I1126 07:24:14.167411 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-content/0.log" Nov 26 07:24:14 crc kubenswrapper[4775]: I1126 07:24:14.378225 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-utilities/0.log" Nov 26 07:24:14 crc kubenswrapper[4775]: I1126 07:24:14.380180 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-content/0.log" Nov 26 07:24:14 crc kubenswrapper[4775]: I1126 07:24:14.912793 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/registry-server/0.log" Nov 26 07:24:21 crc kubenswrapper[4775]: I1126 07:24:21.420229 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:24:21 crc kubenswrapper[4775]: I1126 07:24:21.420760 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.420353 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.420960 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.421042 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.421846 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.421901 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" gracePeriod=600 Nov 26 07:24:51 crc kubenswrapper[4775]: E1126 07:24:51.542248 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.636030 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" exitCode=0 Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.636076 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531"} Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.636110 4775 scope.go:117] "RemoveContainer" containerID="cee8e65f1faefe50945972b8e3690b32b04bca237ff59c2b933c2dacd5b41745" Nov 26 07:24:51 crc kubenswrapper[4775]: I1126 07:24:51.636739 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:24:51 crc kubenswrapper[4775]: E1126 07:24:51.637084 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:25:02 crc kubenswrapper[4775]: I1126 07:25:02.327461 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:25:02 crc kubenswrapper[4775]: E1126 07:25:02.328198 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:25:13 crc kubenswrapper[4775]: I1126 07:25:13.328390 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:25:13 crc kubenswrapper[4775]: E1126 07:25:13.329356 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:25:27 crc kubenswrapper[4775]: I1126 07:25:27.338379 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:25:27 crc kubenswrapper[4775]: E1126 07:25:27.339698 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:25:39 crc kubenswrapper[4775]: I1126 07:25:39.333438 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:25:39 crc kubenswrapper[4775]: E1126 07:25:39.339108 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:25:53 crc kubenswrapper[4775]: I1126 07:25:53.327893 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:25:53 crc kubenswrapper[4775]: E1126 07:25:53.328574 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:25:55 crc kubenswrapper[4775]: I1126 07:25:55.253703 4775 generic.go:334] "Generic (PLEG): container finished" podID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerID="3354087f787034c3409087b94b6d623d122f76079d4afe12d9d82fa27a3f8c1f" exitCode=0 Nov 26 07:25:55 crc kubenswrapper[4775]: I1126 07:25:55.253772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-48sft/must-gather-ljm82" event={"ID":"ad9f4c47-c3af-444f-9b15-48b40663f073","Type":"ContainerDied","Data":"3354087f787034c3409087b94b6d623d122f76079d4afe12d9d82fa27a3f8c1f"} Nov 26 07:25:55 crc kubenswrapper[4775]: I1126 07:25:55.255958 4775 scope.go:117] "RemoveContainer" containerID="3354087f787034c3409087b94b6d623d122f76079d4afe12d9d82fa27a3f8c1f" Nov 26 07:25:55 crc kubenswrapper[4775]: I1126 07:25:55.857095 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-48sft_must-gather-ljm82_ad9f4c47-c3af-444f-9b15-48b40663f073/gather/0.log" Nov 26 07:26:02 crc kubenswrapper[4775]: I1126 07:26:02.960549 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-48sft/must-gather-ljm82"] Nov 26 07:26:02 crc kubenswrapper[4775]: I1126 07:26:02.961549 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-48sft/must-gather-ljm82" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="copy" containerID="cri-o://ffb56d09d07259fc8c523edb26e695b5909d0ad0b85b5d8d9a74991a8f99cd2d" gracePeriod=2 Nov 26 07:26:02 crc kubenswrapper[4775]: I1126 07:26:02.971216 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-48sft/must-gather-ljm82"] Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.368121 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-48sft_must-gather-ljm82_ad9f4c47-c3af-444f-9b15-48b40663f073/copy/0.log" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.370915 4775 generic.go:334] "Generic (PLEG): container finished" podID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerID="ffb56d09d07259fc8c523edb26e695b5909d0ad0b85b5d8d9a74991a8f99cd2d" exitCode=143 Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.370996 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="521e13e0c422f9e3343d2dd6435773df730c9f23694344f3820f8259c2308a0f" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.371414 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-48sft_must-gather-ljm82_ad9f4c47-c3af-444f-9b15-48b40663f073/copy/0.log" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.372428 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.468962 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad9f4c47-c3af-444f-9b15-48b40663f073-must-gather-output\") pod \"ad9f4c47-c3af-444f-9b15-48b40663f073\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.469315 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nz28\" (UniqueName: \"kubernetes.io/projected/ad9f4c47-c3af-444f-9b15-48b40663f073-kube-api-access-8nz28\") pod \"ad9f4c47-c3af-444f-9b15-48b40663f073\" (UID: \"ad9f4c47-c3af-444f-9b15-48b40663f073\") " Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.478479 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9f4c47-c3af-444f-9b15-48b40663f073-kube-api-access-8nz28" (OuterVolumeSpecName: "kube-api-access-8nz28") pod "ad9f4c47-c3af-444f-9b15-48b40663f073" (UID: "ad9f4c47-c3af-444f-9b15-48b40663f073"). InnerVolumeSpecName "kube-api-access-8nz28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.571349 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nz28\" (UniqueName: \"kubernetes.io/projected/ad9f4c47-c3af-444f-9b15-48b40663f073-kube-api-access-8nz28\") on node \"crc\" DevicePath \"\"" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.638234 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad9f4c47-c3af-444f-9b15-48b40663f073-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ad9f4c47-c3af-444f-9b15-48b40663f073" (UID: "ad9f4c47-c3af-444f-9b15-48b40663f073"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:26:03 crc kubenswrapper[4775]: I1126 07:26:03.673684 4775 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ad9f4c47-c3af-444f-9b15-48b40663f073-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 26 07:26:04 crc kubenswrapper[4775]: I1126 07:26:04.327827 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:26:04 crc kubenswrapper[4775]: E1126 07:26:04.328287 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:26:04 crc kubenswrapper[4775]: I1126 07:26:04.378064 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-48sft/must-gather-ljm82" Nov 26 07:26:05 crc kubenswrapper[4775]: I1126 07:26:05.338073 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" path="/var/lib/kubelet/pods/ad9f4c47-c3af-444f-9b15-48b40663f073/volumes" Nov 26 07:26:16 crc kubenswrapper[4775]: I1126 07:26:16.328128 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:26:16 crc kubenswrapper[4775]: E1126 07:26:16.328965 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:26:31 crc kubenswrapper[4775]: I1126 07:26:31.328216 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:26:31 crc kubenswrapper[4775]: E1126 07:26:31.329148 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:26:44 crc kubenswrapper[4775]: I1126 07:26:44.328225 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:26:44 crc kubenswrapper[4775]: E1126 07:26:44.329017 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:26:57 crc kubenswrapper[4775]: I1126 07:26:57.341270 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:26:57 crc kubenswrapper[4775]: E1126 07:26:57.342454 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:27:11 crc kubenswrapper[4775]: I1126 07:27:11.328520 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:27:11 crc kubenswrapper[4775]: E1126 07:27:11.329644 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:27:22 crc kubenswrapper[4775]: I1126 07:27:22.950502 4775 scope.go:117] "RemoveContainer" containerID="3354087f787034c3409087b94b6d623d122f76079d4afe12d9d82fa27a3f8c1f" Nov 26 07:27:23 crc kubenswrapper[4775]: I1126 07:27:23.047452 4775 scope.go:117] "RemoveContainer" containerID="7c51fca7ddf0ab94168303f8f14cb557b25e00703eba689c349b627668c4edaf" Nov 26 07:27:23 crc kubenswrapper[4775]: I1126 07:27:23.076303 4775 scope.go:117] "RemoveContainer" containerID="ffb56d09d07259fc8c523edb26e695b5909d0ad0b85b5d8d9a74991a8f99cd2d" Nov 26 07:27:25 crc kubenswrapper[4775]: I1126 07:27:25.327983 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:27:25 crc kubenswrapper[4775]: E1126 07:27:25.328516 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:27:36 crc kubenswrapper[4775]: I1126 07:27:36.327965 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:27:36 crc kubenswrapper[4775]: E1126 07:27:36.329161 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:27:47 crc kubenswrapper[4775]: I1126 07:27:47.334155 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:27:47 crc kubenswrapper[4775]: E1126 07:27:47.335044 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:27:59 crc kubenswrapper[4775]: I1126 07:27:59.328698 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:27:59 crc kubenswrapper[4775]: E1126 07:27:59.329435 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:28:12 crc kubenswrapper[4775]: I1126 07:28:12.327547 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:28:12 crc kubenswrapper[4775]: E1126 07:28:12.328321 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:28:25 crc kubenswrapper[4775]: I1126 07:28:25.328604 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:28:25 crc kubenswrapper[4775]: E1126 07:28:25.329548 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:28:40 crc kubenswrapper[4775]: I1126 07:28:40.327636 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:28:40 crc kubenswrapper[4775]: E1126 07:28:40.328455 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:28:51 crc kubenswrapper[4775]: I1126 07:28:51.327969 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:28:51 crc kubenswrapper[4775]: E1126 07:28:51.328666 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.307686 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wphn7/must-gather-96w2j"] Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308795 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="extract-utilities" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308815 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="extract-utilities" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308841 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="extract-content" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308850 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="extract-content" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308879 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="gather" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308887 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="gather" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308896 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="registry-server" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308905 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="registry-server" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308924 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="extract-utilities" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308932 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="extract-utilities" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308949 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="copy" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308959 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="copy" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308979 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="registry-server" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.308986 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="registry-server" Nov 26 07:28:59 crc kubenswrapper[4775]: E1126 07:28:59.308998 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="extract-content" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.309005 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="extract-content" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.309225 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b8d6a2-d672-4c1c-9ff6-aba6a2f199c0" containerName="registry-server" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.309236 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="gather" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.309248 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9f4c47-c3af-444f-9b15-48b40663f073" containerName="copy" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.309257 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="572e0013-9c3b-4690-90b2-06eefae2e3cd" containerName="registry-server" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.310480 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.312888 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-wphn7"/"default-dockercfg-8scbm" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.313009 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wphn7"/"kube-root-ca.crt" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.314068 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wphn7"/"openshift-service-ca.crt" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.317080 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wphn7/must-gather-96w2j"] Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.447910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f4rw\" (UniqueName: \"kubernetes.io/projected/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-kube-api-access-7f4rw\") pod \"must-gather-96w2j\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.448216 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-must-gather-output\") pod \"must-gather-96w2j\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.550515 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f4rw\" (UniqueName: \"kubernetes.io/projected/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-kube-api-access-7f4rw\") pod \"must-gather-96w2j\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.550696 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-must-gather-output\") pod \"must-gather-96w2j\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.551188 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-must-gather-output\") pod \"must-gather-96w2j\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.585503 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f4rw\" (UniqueName: \"kubernetes.io/projected/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-kube-api-access-7f4rw\") pod \"must-gather-96w2j\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:28:59 crc kubenswrapper[4775]: I1126 07:28:59.636981 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:29:00 crc kubenswrapper[4775]: I1126 07:29:00.144502 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wphn7/must-gather-96w2j"] Nov 26 07:29:00 crc kubenswrapper[4775]: I1126 07:29:00.990565 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/must-gather-96w2j" event={"ID":"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6","Type":"ContainerStarted","Data":"8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d"} Nov 26 07:29:00 crc kubenswrapper[4775]: I1126 07:29:00.991077 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/must-gather-96w2j" event={"ID":"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6","Type":"ContainerStarted","Data":"880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12"} Nov 26 07:29:00 crc kubenswrapper[4775]: I1126 07:29:00.991092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/must-gather-96w2j" event={"ID":"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6","Type":"ContainerStarted","Data":"063066db080c5560fe7e733da0e55f71a2343498c247d83ed482a07a9e08adbf"} Nov 26 07:29:01 crc kubenswrapper[4775]: I1126 07:29:01.012304 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wphn7/must-gather-96w2j" podStartSLOduration=2.012282132 podStartE2EDuration="2.012282132s" podCreationTimestamp="2025-11-26 07:28:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 07:29:01.008113301 +0000 UTC m=+3824.369417253" watchObservedRunningTime="2025-11-26 07:29:01.012282132 +0000 UTC m=+3824.373586084" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.033612 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wphn7/crc-debug-l74wd"] Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.035591 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.144430 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2g28\" (UniqueName: \"kubernetes.io/projected/6920b1f9-e570-43d3-b101-9f72d1561df1-kube-api-access-g2g28\") pod \"crc-debug-l74wd\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.144916 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6920b1f9-e570-43d3-b101-9f72d1561df1-host\") pod \"crc-debug-l74wd\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.247447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2g28\" (UniqueName: \"kubernetes.io/projected/6920b1f9-e570-43d3-b101-9f72d1561df1-kube-api-access-g2g28\") pod \"crc-debug-l74wd\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.247603 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6920b1f9-e570-43d3-b101-9f72d1561df1-host\") pod \"crc-debug-l74wd\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.247796 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6920b1f9-e570-43d3-b101-9f72d1561df1-host\") pod \"crc-debug-l74wd\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.275458 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2g28\" (UniqueName: \"kubernetes.io/projected/6920b1f9-e570-43d3-b101-9f72d1561df1-kube-api-access-g2g28\") pod \"crc-debug-l74wd\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:04 crc kubenswrapper[4775]: I1126 07:29:04.357869 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:05 crc kubenswrapper[4775]: I1126 07:29:05.030021 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-l74wd" event={"ID":"6920b1f9-e570-43d3-b101-9f72d1561df1","Type":"ContainerStarted","Data":"b2eabb09df6007c8a906def493b69a0b85457228e2088b58fb98affe76970e81"} Nov 26 07:29:05 crc kubenswrapper[4775]: I1126 07:29:05.030570 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-l74wd" event={"ID":"6920b1f9-e570-43d3-b101-9f72d1561df1","Type":"ContainerStarted","Data":"fb5aa0e34cb658820976a2c1dcb04151228f7d70aa36942b6482703164eef6e1"} Nov 26 07:29:05 crc kubenswrapper[4775]: I1126 07:29:05.054889 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wphn7/crc-debug-l74wd" podStartSLOduration=1.054868271 podStartE2EDuration="1.054868271s" podCreationTimestamp="2025-11-26 07:29:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 07:29:05.049887307 +0000 UTC m=+3828.411191259" watchObservedRunningTime="2025-11-26 07:29:05.054868271 +0000 UTC m=+3828.416172233" Nov 26 07:29:06 crc kubenswrapper[4775]: I1126 07:29:06.327613 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:29:06 crc kubenswrapper[4775]: E1126 07:29:06.328929 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:29:18 crc kubenswrapper[4775]: I1126 07:29:18.328461 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:29:18 crc kubenswrapper[4775]: E1126 07:29:18.329179 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:29:29 crc kubenswrapper[4775]: I1126 07:29:29.327740 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:29:29 crc kubenswrapper[4775]: E1126 07:29:29.328552 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:29:39 crc kubenswrapper[4775]: I1126 07:29:39.346870 4775 generic.go:334] "Generic (PLEG): container finished" podID="6920b1f9-e570-43d3-b101-9f72d1561df1" containerID="b2eabb09df6007c8a906def493b69a0b85457228e2088b58fb98affe76970e81" exitCode=0 Nov 26 07:29:39 crc kubenswrapper[4775]: I1126 07:29:39.346967 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-l74wd" event={"ID":"6920b1f9-e570-43d3-b101-9f72d1561df1","Type":"ContainerDied","Data":"b2eabb09df6007c8a906def493b69a0b85457228e2088b58fb98affe76970e81"} Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.453359 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.492743 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wphn7/crc-debug-l74wd"] Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.505135 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wphn7/crc-debug-l74wd"] Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.507534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2g28\" (UniqueName: \"kubernetes.io/projected/6920b1f9-e570-43d3-b101-9f72d1561df1-kube-api-access-g2g28\") pod \"6920b1f9-e570-43d3-b101-9f72d1561df1\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.507618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6920b1f9-e570-43d3-b101-9f72d1561df1-host\") pod \"6920b1f9-e570-43d3-b101-9f72d1561df1\" (UID: \"6920b1f9-e570-43d3-b101-9f72d1561df1\") " Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.508548 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6920b1f9-e570-43d3-b101-9f72d1561df1-host" (OuterVolumeSpecName: "host") pod "6920b1f9-e570-43d3-b101-9f72d1561df1" (UID: "6920b1f9-e570-43d3-b101-9f72d1561df1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.513668 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6920b1f9-e570-43d3-b101-9f72d1561df1-kube-api-access-g2g28" (OuterVolumeSpecName: "kube-api-access-g2g28") pod "6920b1f9-e570-43d3-b101-9f72d1561df1" (UID: "6920b1f9-e570-43d3-b101-9f72d1561df1"). InnerVolumeSpecName "kube-api-access-g2g28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.610690 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2g28\" (UniqueName: \"kubernetes.io/projected/6920b1f9-e570-43d3-b101-9f72d1561df1-kube-api-access-g2g28\") on node \"crc\" DevicePath \"\"" Nov 26 07:29:40 crc kubenswrapper[4775]: I1126 07:29:40.610774 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6920b1f9-e570-43d3-b101-9f72d1561df1-host\") on node \"crc\" DevicePath \"\"" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.342009 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6920b1f9-e570-43d3-b101-9f72d1561df1" path="/var/lib/kubelet/pods/6920b1f9-e570-43d3-b101-9f72d1561df1/volumes" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.365917 4775 scope.go:117] "RemoveContainer" containerID="b2eabb09df6007c8a906def493b69a0b85457228e2088b58fb98affe76970e81" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.366141 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-l74wd" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.682242 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wphn7/crc-debug-s5bk7"] Nov 26 07:29:41 crc kubenswrapper[4775]: E1126 07:29:41.682664 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6920b1f9-e570-43d3-b101-9f72d1561df1" containerName="container-00" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.682677 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6920b1f9-e570-43d3-b101-9f72d1561df1" containerName="container-00" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.682881 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6920b1f9-e570-43d3-b101-9f72d1561df1" containerName="container-00" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.683478 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.731379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7g2d\" (UniqueName: \"kubernetes.io/projected/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-kube-api-access-d7g2d\") pod \"crc-debug-s5bk7\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.731568 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-host\") pod \"crc-debug-s5bk7\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.832882 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-host\") pod \"crc-debug-s5bk7\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.833036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-host\") pod \"crc-debug-s5bk7\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.833052 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7g2d\" (UniqueName: \"kubernetes.io/projected/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-kube-api-access-d7g2d\") pod \"crc-debug-s5bk7\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:41 crc kubenswrapper[4775]: I1126 07:29:41.853835 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7g2d\" (UniqueName: \"kubernetes.io/projected/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-kube-api-access-d7g2d\") pod \"crc-debug-s5bk7\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:42 crc kubenswrapper[4775]: I1126 07:29:42.001947 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:42 crc kubenswrapper[4775]: I1126 07:29:42.378657 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab1931c1-fb38-4e2b-b4ec-2f2671c15231" containerID="d2b4c118bf583f664f6845d59acf19d4481cdd3bafae2c6858d672e3f73723da" exitCode=0 Nov 26 07:29:42 crc kubenswrapper[4775]: I1126 07:29:42.378849 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-s5bk7" event={"ID":"ab1931c1-fb38-4e2b-b4ec-2f2671c15231","Type":"ContainerDied","Data":"d2b4c118bf583f664f6845d59acf19d4481cdd3bafae2c6858d672e3f73723da"} Nov 26 07:29:42 crc kubenswrapper[4775]: I1126 07:29:42.379298 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-s5bk7" event={"ID":"ab1931c1-fb38-4e2b-b4ec-2f2671c15231","Type":"ContainerStarted","Data":"544707dc9211a7637af046500e7bd127457cc58cc297321f60451a7a8dc52d24"} Nov 26 07:29:42 crc kubenswrapper[4775]: I1126 07:29:42.864359 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wphn7/crc-debug-s5bk7"] Nov 26 07:29:42 crc kubenswrapper[4775]: I1126 07:29:42.885008 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wphn7/crc-debug-s5bk7"] Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.495252 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.666986 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-host\") pod \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.667126 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-host" (OuterVolumeSpecName: "host") pod "ab1931c1-fb38-4e2b-b4ec-2f2671c15231" (UID: "ab1931c1-fb38-4e2b-b4ec-2f2671c15231"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.667263 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7g2d\" (UniqueName: \"kubernetes.io/projected/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-kube-api-access-d7g2d\") pod \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\" (UID: \"ab1931c1-fb38-4e2b-b4ec-2f2671c15231\") " Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.667790 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-host\") on node \"crc\" DevicePath \"\"" Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.676556 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-kube-api-access-d7g2d" (OuterVolumeSpecName: "kube-api-access-d7g2d") pod "ab1931c1-fb38-4e2b-b4ec-2f2671c15231" (UID: "ab1931c1-fb38-4e2b-b4ec-2f2671c15231"). InnerVolumeSpecName "kube-api-access-d7g2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:29:43 crc kubenswrapper[4775]: I1126 07:29:43.769214 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7g2d\" (UniqueName: \"kubernetes.io/projected/ab1931c1-fb38-4e2b-b4ec-2f2671c15231-kube-api-access-d7g2d\") on node \"crc\" DevicePath \"\"" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.073577 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wphn7/crc-debug-nwglz"] Nov 26 07:29:44 crc kubenswrapper[4775]: E1126 07:29:44.074385 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1931c1-fb38-4e2b-b4ec-2f2671c15231" containerName="container-00" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.074407 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1931c1-fb38-4e2b-b4ec-2f2671c15231" containerName="container-00" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.074764 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab1931c1-fb38-4e2b-b4ec-2f2671c15231" containerName="container-00" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.075877 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.176365 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2tnm\" (UniqueName: \"kubernetes.io/projected/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-kube-api-access-l2tnm\") pod \"crc-debug-nwglz\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.176872 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-host\") pod \"crc-debug-nwglz\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.278276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2tnm\" (UniqueName: \"kubernetes.io/projected/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-kube-api-access-l2tnm\") pod \"crc-debug-nwglz\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.278373 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-host\") pod \"crc-debug-nwglz\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.278476 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-host\") pod \"crc-debug-nwglz\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.315071 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2tnm\" (UniqueName: \"kubernetes.io/projected/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-kube-api-access-l2tnm\") pod \"crc-debug-nwglz\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.328492 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:29:44 crc kubenswrapper[4775]: E1126 07:29:44.328769 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.394223 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.402981 4775 scope.go:117] "RemoveContainer" containerID="d2b4c118bf583f664f6845d59acf19d4481cdd3bafae2c6858d672e3f73723da" Nov 26 07:29:44 crc kubenswrapper[4775]: I1126 07:29:44.403013 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-s5bk7" Nov 26 07:29:44 crc kubenswrapper[4775]: W1126 07:29:44.444465 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc64623d_18b1_4cbe_a50f_5b8fc5cfe450.slice/crio-d980b34412d148b8bccb4c97b047c71e297de4f932a974271fc294fdf2aaf442 WatchSource:0}: Error finding container d980b34412d148b8bccb4c97b047c71e297de4f932a974271fc294fdf2aaf442: Status 404 returned error can't find the container with id d980b34412d148b8bccb4c97b047c71e297de4f932a974271fc294fdf2aaf442 Nov 26 07:29:45 crc kubenswrapper[4775]: I1126 07:29:45.338480 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab1931c1-fb38-4e2b-b4ec-2f2671c15231" path="/var/lib/kubelet/pods/ab1931c1-fb38-4e2b-b4ec-2f2671c15231/volumes" Nov 26 07:29:45 crc kubenswrapper[4775]: I1126 07:29:45.414579 4775 generic.go:334] "Generic (PLEG): container finished" podID="fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" containerID="ef89959f9bdae38302d7e9daf6b64e63b480e86a9d508eeb5c4a09705f0978e7" exitCode=0 Nov 26 07:29:45 crc kubenswrapper[4775]: I1126 07:29:45.414631 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-nwglz" event={"ID":"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450","Type":"ContainerDied","Data":"ef89959f9bdae38302d7e9daf6b64e63b480e86a9d508eeb5c4a09705f0978e7"} Nov 26 07:29:45 crc kubenswrapper[4775]: I1126 07:29:45.414660 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/crc-debug-nwglz" event={"ID":"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450","Type":"ContainerStarted","Data":"d980b34412d148b8bccb4c97b047c71e297de4f932a974271fc294fdf2aaf442"} Nov 26 07:29:45 crc kubenswrapper[4775]: I1126 07:29:45.449387 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wphn7/crc-debug-nwglz"] Nov 26 07:29:45 crc kubenswrapper[4775]: I1126 07:29:45.458378 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wphn7/crc-debug-nwglz"] Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.533079 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.718908 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2tnm\" (UniqueName: \"kubernetes.io/projected/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-kube-api-access-l2tnm\") pod \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.719228 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-host\") pod \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\" (UID: \"fc64623d-18b1-4cbe-a50f-5b8fc5cfe450\") " Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.719486 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-host" (OuterVolumeSpecName: "host") pod "fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" (UID: "fc64623d-18b1-4cbe-a50f-5b8fc5cfe450"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.720178 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-host\") on node \"crc\" DevicePath \"\"" Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.726988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-kube-api-access-l2tnm" (OuterVolumeSpecName: "kube-api-access-l2tnm") pod "fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" (UID: "fc64623d-18b1-4cbe-a50f-5b8fc5cfe450"). InnerVolumeSpecName "kube-api-access-l2tnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:29:46 crc kubenswrapper[4775]: I1126 07:29:46.821421 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2tnm\" (UniqueName: \"kubernetes.io/projected/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450-kube-api-access-l2tnm\") on node \"crc\" DevicePath \"\"" Nov 26 07:29:47 crc kubenswrapper[4775]: I1126 07:29:47.340039 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" path="/var/lib/kubelet/pods/fc64623d-18b1-4cbe-a50f-5b8fc5cfe450/volumes" Nov 26 07:29:47 crc kubenswrapper[4775]: I1126 07:29:47.442968 4775 scope.go:117] "RemoveContainer" containerID="ef89959f9bdae38302d7e9daf6b64e63b480e86a9d508eeb5c4a09705f0978e7" Nov 26 07:29:47 crc kubenswrapper[4775]: I1126 07:29:47.442990 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/crc-debug-nwglz" Nov 26 07:29:59 crc kubenswrapper[4775]: I1126 07:29:59.327447 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.207839 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc"] Nov 26 07:30:00 crc kubenswrapper[4775]: E1126 07:30:00.208881 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" containerName="container-00" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.208904 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" containerName="container-00" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.209161 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc64623d-18b1-4cbe-a50f-5b8fc5cfe450" containerName="container-00" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.210002 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.212328 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.212587 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.216193 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc"] Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.379779 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvtqx\" (UniqueName: \"kubernetes.io/projected/b13da712-45ab-43fa-9435-38a7f111f60a-kube-api-access-rvtqx\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.384487 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b13da712-45ab-43fa-9435-38a7f111f60a-secret-volume\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.384978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b13da712-45ab-43fa-9435-38a7f111f60a-config-volume\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.486811 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b13da712-45ab-43fa-9435-38a7f111f60a-secret-volume\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.486910 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b13da712-45ab-43fa-9435-38a7f111f60a-config-volume\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.486956 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvtqx\" (UniqueName: \"kubernetes.io/projected/b13da712-45ab-43fa-9435-38a7f111f60a-kube-api-access-rvtqx\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.488328 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b13da712-45ab-43fa-9435-38a7f111f60a-config-volume\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.493270 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b13da712-45ab-43fa-9435-38a7f111f60a-secret-volume\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.514537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvtqx\" (UniqueName: \"kubernetes.io/projected/b13da712-45ab-43fa-9435-38a7f111f60a-kube-api-access-rvtqx\") pod \"collect-profiles-29402370-vckrc\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.546394 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:00 crc kubenswrapper[4775]: I1126 07:30:00.568579 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"988d3d41547c245627cb395ac65ae477aa64ca8ebb71a5826ab044045474b18e"} Nov 26 07:30:01 crc kubenswrapper[4775]: I1126 07:30:01.051372 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc"] Nov 26 07:30:01 crc kubenswrapper[4775]: W1126 07:30:01.053620 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb13da712_45ab_43fa_9435_38a7f111f60a.slice/crio-1ba768ba2c91820e79fec32aa9d56d23184e88dc2f668e6cb077b67ccd9cd81e WatchSource:0}: Error finding container 1ba768ba2c91820e79fec32aa9d56d23184e88dc2f668e6cb077b67ccd9cd81e: Status 404 returned error can't find the container with id 1ba768ba2c91820e79fec32aa9d56d23184e88dc2f668e6cb077b67ccd9cd81e Nov 26 07:30:01 crc kubenswrapper[4775]: I1126 07:30:01.577481 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" event={"ID":"b13da712-45ab-43fa-9435-38a7f111f60a","Type":"ContainerStarted","Data":"b987b2be23e27ac65551a402ee9cd771a6964a5ef09f1dc1bc7f48b5244689c3"} Nov 26 07:30:01 crc kubenswrapper[4775]: I1126 07:30:01.577831 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" event={"ID":"b13da712-45ab-43fa-9435-38a7f111f60a","Type":"ContainerStarted","Data":"1ba768ba2c91820e79fec32aa9d56d23184e88dc2f668e6cb077b67ccd9cd81e"} Nov 26 07:30:01 crc kubenswrapper[4775]: I1126 07:30:01.593986 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" podStartSLOduration=1.5939624700000001 podStartE2EDuration="1.59396247s" podCreationTimestamp="2025-11-26 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 07:30:01.592266685 +0000 UTC m=+3884.953570667" watchObservedRunningTime="2025-11-26 07:30:01.59396247 +0000 UTC m=+3884.955266432" Nov 26 07:30:02 crc kubenswrapper[4775]: I1126 07:30:02.587528 4775 generic.go:334] "Generic (PLEG): container finished" podID="b13da712-45ab-43fa-9435-38a7f111f60a" containerID="b987b2be23e27ac65551a402ee9cd771a6964a5ef09f1dc1bc7f48b5244689c3" exitCode=0 Nov 26 07:30:02 crc kubenswrapper[4775]: I1126 07:30:02.587582 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" event={"ID":"b13da712-45ab-43fa-9435-38a7f111f60a","Type":"ContainerDied","Data":"b987b2be23e27ac65551a402ee9cd771a6964a5ef09f1dc1bc7f48b5244689c3"} Nov 26 07:30:03 crc kubenswrapper[4775]: I1126 07:30:03.971358 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.165099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvtqx\" (UniqueName: \"kubernetes.io/projected/b13da712-45ab-43fa-9435-38a7f111f60a-kube-api-access-rvtqx\") pod \"b13da712-45ab-43fa-9435-38a7f111f60a\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.165619 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b13da712-45ab-43fa-9435-38a7f111f60a-config-volume\") pod \"b13da712-45ab-43fa-9435-38a7f111f60a\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.165796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b13da712-45ab-43fa-9435-38a7f111f60a-secret-volume\") pod \"b13da712-45ab-43fa-9435-38a7f111f60a\" (UID: \"b13da712-45ab-43fa-9435-38a7f111f60a\") " Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.166484 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13da712-45ab-43fa-9435-38a7f111f60a-config-volume" (OuterVolumeSpecName: "config-volume") pod "b13da712-45ab-43fa-9435-38a7f111f60a" (UID: "b13da712-45ab-43fa-9435-38a7f111f60a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.171336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13da712-45ab-43fa-9435-38a7f111f60a-kube-api-access-rvtqx" (OuterVolumeSpecName: "kube-api-access-rvtqx") pod "b13da712-45ab-43fa-9435-38a7f111f60a" (UID: "b13da712-45ab-43fa-9435-38a7f111f60a"). InnerVolumeSpecName "kube-api-access-rvtqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.171648 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13da712-45ab-43fa-9435-38a7f111f60a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b13da712-45ab-43fa-9435-38a7f111f60a" (UID: "b13da712-45ab-43fa-9435-38a7f111f60a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.267845 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvtqx\" (UniqueName: \"kubernetes.io/projected/b13da712-45ab-43fa-9435-38a7f111f60a-kube-api-access-rvtqx\") on node \"crc\" DevicePath \"\"" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.267916 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b13da712-45ab-43fa-9435-38a7f111f60a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.267929 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b13da712-45ab-43fa-9435-38a7f111f60a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.609421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" event={"ID":"b13da712-45ab-43fa-9435-38a7f111f60a","Type":"ContainerDied","Data":"1ba768ba2c91820e79fec32aa9d56d23184e88dc2f668e6cb077b67ccd9cd81e"} Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.609506 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ba768ba2c91820e79fec32aa9d56d23184e88dc2f668e6cb077b67ccd9cd81e" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.609571 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402370-vckrc" Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.671182 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9"] Nov 26 07:30:04 crc kubenswrapper[4775]: I1126 07:30:04.685476 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402325-mkjf9"] Nov 26 07:30:05 crc kubenswrapper[4775]: I1126 07:30:05.349702 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86378490-4d35-4d6f-8b0e-16225f624014" path="/var/lib/kubelet/pods/86378490-4d35-4d6f-8b0e-16225f624014/volumes" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.226190 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd8c567d4-z5vr4_1db6f2a6-c3be-4317-a6ce-802242830778/barbican-api/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.356356 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fd8c567d4-z5vr4_1db6f2a6-c3be-4317-a6ce-802242830778/barbican-api-log/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.432533 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-76d59b886b-lsps8_fa465504-4f8f-4c22-8f6b-3bb834f303c2/barbican-keystone-listener/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.511643 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-76d59b886b-lsps8_fa465504-4f8f-4c22-8f6b-3bb834f303c2/barbican-keystone-listener-log/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.622626 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bdbc86c75-sgb2x_2b958489-5435-42a4-bff8-577a26c717c0/barbican-worker-log/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.633210 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5bdbc86c75-sgb2x_2b958489-5435-42a4-bff8-577a26c717c0/barbican-worker/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.775599 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-lsdzz_66abe898-3799-4e9d-abc7-748423d2daa3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.851479 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/ceilometer-central-agent/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.918970 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/ceilometer-notification-agent/0.log" Nov 26 07:30:10 crc kubenswrapper[4775]: I1126 07:30:10.956599 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/proxy-httpd/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.031633 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_9bf2ed3a-41d4-4ca1-8b5c-4e4e85b0ebd7/sg-core/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.153172 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b9daeb8-621b-4557-9757-6e8b9c430339/cinder-api-log/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.188282 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_4b9daeb8-621b-4557-9757-6e8b9c430339/cinder-api/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.417960 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3bc9a6f7-ee1c-49c1-855d-362fa796cf07/probe/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.433629 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3bc9a6f7-ee1c-49c1-855d-362fa796cf07/cinder-scheduler/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.556123 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-kp72k_7b826be9-41b3-4715-b7cb-5b8bc10e35a9/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.606990 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-kmx5g_c6dbfaeb-21ee-4017-aab7-6d1219d7e10f/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.752869 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-dhkrg_e837eeaf-6cf5-44e0-b183-334f62304bc7/init/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.927650 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-dhkrg_e837eeaf-6cf5-44e0-b183-334f62304bc7/init/0.log" Nov 26 07:30:11 crc kubenswrapper[4775]: I1126 07:30:11.969376 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-dhkrg_e837eeaf-6cf5-44e0-b183-334f62304bc7/dnsmasq-dns/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.008543 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-6zbsh_e780accc-4808-46d6-b5df-5b6492107f25/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.253069 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58818a45-1408-4416-b18e-814acd04b059/glance-httpd/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.436874 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_46819d15-5e66-4501-b2f0-570f3b768207/glance-httpd/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.453434 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58818a45-1408-4416-b18e-814acd04b059/glance-log/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.483776 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_46819d15-5e66-4501-b2f0-570f3b768207/glance-log/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.686874 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-9c9599df6-zd8gh_f578a3d5-f846-44b6-bc2b-3d98188dffff/horizon/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.922740 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-svnzx_0cad897c-5d2b-496c-8e45-e5061ca93a6a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:12 crc kubenswrapper[4775]: I1126 07:30:12.935826 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-dndkl_22cadc11-2c9d-49f4-abf7-2e6fa3f69905/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.061225 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-9c9599df6-zd8gh_f578a3d5-f846-44b6-bc2b-3d98188dffff/horizon-log/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.248115 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7679ff5658-lwrpm_f06e8ee2-5844-487f-80e7-cdbba5909f74/keystone-api/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.321392 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29402341-lzbf6_55d5f81a-1fab-4bfa-8ca9-115de4fef547/keystone-cron/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.479705 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2427f2bd-7928-4ff9-bc65-23bc48225184/kube-state-metrics/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.548203 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-zrq4j_da48dd91-7390-4751-8e73-455fda656b0d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.905649 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54d985cc65-zsv4d_e33bc539-83a2-4077-a430-4872e8587023/neutron-api/0.log" Nov 26 07:30:13 crc kubenswrapper[4775]: I1126 07:30:13.995806 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54d985cc65-zsv4d_e33bc539-83a2-4077-a430-4872e8587023/neutron-httpd/0.log" Nov 26 07:30:14 crc kubenswrapper[4775]: I1126 07:30:14.136685 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-qmx27_51e229f8-0c28-4b0c-8307-a2dc128fa503/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:14 crc kubenswrapper[4775]: I1126 07:30:14.766464 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_09a96730-d166-4111-b88a-b2a51ca06b5a/nova-api-log/0.log" Nov 26 07:30:14 crc kubenswrapper[4775]: I1126 07:30:14.850802 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_76cb0fc1-4a71-4ae4-9b2e-7483efe206d7/nova-cell0-conductor-conductor/0.log" Nov 26 07:30:15 crc kubenswrapper[4775]: I1126 07:30:15.059150 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_09a96730-d166-4111-b88a-b2a51ca06b5a/nova-api-api/0.log" Nov 26 07:30:15 crc kubenswrapper[4775]: I1126 07:30:15.767122 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c7b39d9f-bad3-4e37-b5b0-f5aaa6b7c211/nova-cell1-novncproxy-novncproxy/0.log" Nov 26 07:30:15 crc kubenswrapper[4775]: I1126 07:30:15.819696 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ec4083e2-96ad-420c-a66f-0cb0f274867a/nova-cell1-conductor-conductor/0.log" Nov 26 07:30:16 crc kubenswrapper[4775]: I1126 07:30:16.009281 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-ldl57_fc8364bf-1ce5-4681-99a5-edc754f142ac/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:16 crc kubenswrapper[4775]: I1126 07:30:16.149437 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2d57ebfc-a725-46a3-9224-f011eadfc783/nova-metadata-log/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.108351 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e8d44208-4c24-4835-89f6-a471ce50ddae/nova-scheduler-scheduler/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.116850 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c/mysql-bootstrap/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.368223 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c/mysql-bootstrap/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.421603 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7aa3f7db-fcc8-4c33-93eb-d75d66c3a81c/galera/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.725101 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0028de9f-a7db-47c0-a552-e9f518435458/mysql-bootstrap/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.904609 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0028de9f-a7db-47c0-a552-e9f518435458/mysql-bootstrap/0.log" Nov 26 07:30:17 crc kubenswrapper[4775]: I1126 07:30:17.933570 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0028de9f-a7db-47c0-a552-e9f518435458/galera/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.070005 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_2d57ebfc-a725-46a3-9224-f011eadfc783/nova-metadata-metadata/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.169485 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-k4mqj_d30382f9-28bd-4ab7-8fec-1f8932b3f804/ovn-controller/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.173070 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3675965a-9ca7-433b-9c69-99596e74aa25/openstackclient/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.368602 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovsdb-server-init/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.410173 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-kd448_ff2f6960-e698-4a01-9a6d-9b1fd470e295/openstack-network-exporter/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.769660 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovsdb-server/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.783475 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovs-vswitchd/0.log" Nov 26 07:30:18 crc kubenswrapper[4775]: I1126 07:30:18.801398 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4qmsw_64dca619-067a-4f6d-b391-c736d9771a1a/ovsdb-server-init/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.007332 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gfrt7_62bfa377-ebb1-48d4-82b6-eacd0ac68c47/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.018910 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2f5d800-59f5-41fa-b45b-f2fa0981b3da/openstack-network-exporter/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.041669 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c2f5d800-59f5-41fa-b45b-f2fa0981b3da/ovn-northd/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.193469 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e61da90a-7acf-4b95-934b-69f09efb5cbc/openstack-network-exporter/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.253266 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_e61da90a-7acf-4b95-934b-69f09efb5cbc/ovsdbserver-nb/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.440488 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dc88dc0e-14a1-479e-9682-bb5c611259a8/openstack-network-exporter/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.464974 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dc88dc0e-14a1-479e-9682-bb5c611259a8/ovsdbserver-sb/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.684654 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd45b998d-gjx5d_bc391c06-ad32-4c18-9670-481f7614e1a3/placement-api/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.799326 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ce90c242-2ac3-4509-bf08-2f7c77b1aff0/setup-container/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.865029 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cd45b998d-gjx5d_bc391c06-ad32-4c18-9670-481f7614e1a3/placement-log/0.log" Nov 26 07:30:19 crc kubenswrapper[4775]: I1126 07:30:19.938879 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ce90c242-2ac3-4509-bf08-2f7c77b1aff0/setup-container/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.069599 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_ce90c242-2ac3-4509-bf08-2f7c77b1aff0/rabbitmq/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.105182 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_837ab270-d963-4406-9a55-390bf0611e14/setup-container/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.306633 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_837ab270-d963-4406-9a55-390bf0611e14/setup-container/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.323108 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-dbvxl_efef5423-9419-4ff8-b603-96ffaa7f4dcd/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.335172 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_837ab270-d963-4406-9a55-390bf0611e14/rabbitmq/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.528593 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-gm5w2_f949db3e-70cb-458b-96d9-9c9f0f1118d8/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.624029 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5cf57_561d8a6c-3d90-4220-b6ac-386d81f5016f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.742950 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-lljnf_6233c3cd-e8bc-4dd1-b039-c950e119ce4e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:20 crc kubenswrapper[4775]: I1126 07:30:20.904606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ghkc9_7e9b9292-e385-4016-8355-2b95cf0e85af/ssh-known-hosts-edpm-deployment/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.061541 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6c8b9f5bf9-nsz2f_546c7a7f-b54b-415e-81c6-0fb3fa734e05/proxy-server/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.214853 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6c8b9f5bf9-nsz2f_546c7a7f-b54b-415e-81c6-0fb3fa734e05/proxy-httpd/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.242606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-d8msj_fd3c60bd-b664-43b1-b669-3ef640959006/swift-ring-rebalance/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.377152 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-auditor/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.444439 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-reaper/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.491589 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-replicator/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.580952 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/account-server/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.632244 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-auditor/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.710682 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-server/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.742879 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-replicator/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.792927 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/container-updater/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.876409 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-auditor/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.922051 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-expirer/0.log" Nov 26 07:30:21 crc kubenswrapper[4775]: I1126 07:30:21.954811 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-replicator/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.007359 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-server/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.088338 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/object-updater/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.155003 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/rsync/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.185163 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_38df1b80-132c-467f-969c-d3798dd1f7d9/swift-recon-cron/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.404946 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2ssps_1ad4c3cd-bb52-4245-b8f9-4c6bd20d3dcf/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.413617 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_da6eaea0-ca0d-43ba-b2e7-9d61a19f1cc5/tempest-tests-tempest-tests-runner/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.572902 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_7075bbd0-afa7-4430-b2b9-ee07eac73b2a/test-operator-logs-container/0.log" Nov 26 07:30:22 crc kubenswrapper[4775]: I1126 07:30:22.684348 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-z74zq_3d9ee8c3-9727-4454-8df5-5ecce45daf45/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 26 07:30:23 crc kubenswrapper[4775]: I1126 07:30:23.187255 4775 scope.go:117] "RemoveContainer" containerID="19833754e9909e7e7724dd83c173f935a074c4d511545d0452cfbb02fa8b3835" Nov 26 07:30:31 crc kubenswrapper[4775]: I1126 07:30:31.504562 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_bf022b43-4760-4e21-8f5b-2558a69ede14/memcached/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.272605 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/util/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.463267 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/pull/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.471141 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/pull/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.485457 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/util/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.657497 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/extract/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.673089 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/pull/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.713865 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_83184424b226d06e630cd19771778c95a5aec6cd84d7bf7524feedbe9etkcff_78b0fc74-36dd-4791-bc5c-dd6b2d6b27bb/util/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.852640 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-27cqk_4f5effa7-3458-4978-8382-0f5b4a17105a/kube-rbac-proxy/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.882043 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7b64f4fb85-27cqk_4f5effa7-3458-4978-8382-0f5b4a17105a/manager/0.log" Nov 26 07:30:46 crc kubenswrapper[4775]: I1126 07:30:46.929329 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-5whjf_153b27e4-d948-4ff6-9d6a-bfbcebd17cad/kube-rbac-proxy/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.069034 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6b7f75547b-5whjf_153b27e4-d948-4ff6-9d6a-bfbcebd17cad/manager/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.112049 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dghfx_1a1d393e-2d6c-425c-9e6f-068d4c8ee090/kube-rbac-proxy/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.139978 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-955677c94-dghfx_1a1d393e-2d6c-425c-9e6f-068d4c8ee090/manager/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.287004 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-sdz7q_cc351f29-006f-4ee4-80cb-5860f4f824ff/kube-rbac-proxy/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.356202 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-589cbd6b5b-sdz7q_cc351f29-006f-4ee4-80cb-5860f4f824ff/manager/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.459906 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-d7m5d_faa5a2e8-ca2b-42e1-ac9a-3580111707ac/kube-rbac-proxy/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.525661 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5b77f656f-d7m5d_faa5a2e8-ca2b-42e1-ac9a-3580111707ac/manager/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.648546 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-jmwr9_ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7/kube-rbac-proxy/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.664361 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5d494799bf-jmwr9_ca53f3d4-1c70-49b4-8aa2-dd65f2659fb7/manager/0.log" Nov 26 07:30:47 crc kubenswrapper[4775]: I1126 07:30:47.761120 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-xrzlh_8483a9f8-d70d-497d-ae02-c6b426a49306/kube-rbac-proxy/0.log" Nov 26 07:30:48 crc kubenswrapper[4775]: I1126 07:30:48.373179 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-gdk4v_2158b3e8-3c08-4e7c-9b28-cd32534135a1/kube-rbac-proxy/0.log" Nov 26 07:30:48 crc kubenswrapper[4775]: I1126 07:30:48.432108 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67cb4dc6d4-gdk4v_2158b3e8-3c08-4e7c-9b28-cd32534135a1/manager/0.log" Nov 26 07:30:48 crc kubenswrapper[4775]: I1126 07:30:48.493070 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-xrzlh_8483a9f8-d70d-497d-ae02-c6b426a49306/manager/0.log" Nov 26 07:30:48 crc kubenswrapper[4775]: I1126 07:30:48.636969 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4fxcp_997d00b8-4cc3-45e9-8af2-2d573eef844a/kube-rbac-proxy/0.log" Nov 26 07:30:48 crc kubenswrapper[4775]: I1126 07:30:48.763240 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b4567c7cf-4fxcp_997d00b8-4cc3-45e9-8af2-2d573eef844a/manager/0.log" Nov 26 07:30:48 crc kubenswrapper[4775]: I1126 07:30:48.851212 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-5hvj7_021e1511-a060-4d30-af62-872ba84b84c3/kube-rbac-proxy/0.log" Nov 26 07:30:49 crc kubenswrapper[4775]: I1126 07:30:49.196836 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-6jrgg_85c90cee-6304-4ab9-957c-0bd91411403f/kube-rbac-proxy/0.log" Nov 26 07:30:49 crc kubenswrapper[4775]: I1126 07:30:49.229545 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5d499bf58b-5hvj7_021e1511-a060-4d30-af62-872ba84b84c3/manager/0.log" Nov 26 07:30:49 crc kubenswrapper[4775]: I1126 07:30:49.310078 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-66f4dd4bc7-6jrgg_85c90cee-6304-4ab9-957c-0bd91411403f/manager/0.log" Nov 26 07:30:49 crc kubenswrapper[4775]: I1126 07:30:49.501526 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-q97hn_7bead1ee-7a42-45c9-9205-420ae85002f7/manager/0.log" Nov 26 07:30:49 crc kubenswrapper[4775]: I1126 07:30:49.505785 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6fdcddb789-q97hn_7bead1ee-7a42-45c9-9205-420ae85002f7/kube-rbac-proxy/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.113618 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ftsgz_d8ad3c0b-931e-444d-9fcf-23dfbe77bedc/kube-rbac-proxy/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.137988 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-plpx4_d22fe1a8-9183-4ca4-ae47-60d3e77f2339/kube-rbac-proxy/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.218548 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-ftsgz_d8ad3c0b-931e-444d-9fcf-23dfbe77bedc/manager/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.342369 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-64cdc6ff96-plpx4_d22fe1a8-9183-4ca4-ae47-60d3e77f2339/manager/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.376064 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv_7ac95117-0ba0-4c51-afbf-320322b8ebfc/kube-rbac-proxy/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.447761 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5fcdb54b6bfslpv_7ac95117-0ba0-4c51-afbf-320322b8ebfc/manager/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.785884 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6df95dcb7f-5wnrd_2b412dfe-90b0-4018-81ca-32d9644fbc99/operator/0.log" Nov 26 07:30:50 crc kubenswrapper[4775]: I1126 07:30:50.804468 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rpzlv_cd5862a0-b9ab-4473-bbe4-34253c8cbb79/registry-server/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.036116 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-ksct6_5e867f5f-7847-4469-95cb-605042211f56/kube-rbac-proxy/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.086608 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-56897c768d-ksct6_5e867f5f-7847-4469-95cb-605042211f56/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.094056 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-ffw7r_f9b85bd6-21c3-457e-81c0-b87c56be9f3e/kube-rbac-proxy/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.250256 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-57988cc5b5-ffw7r_f9b85bd6-21c3-457e-81c0-b87c56be9f3e/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.327914 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-r7dqc_2e76307b-7079-41d7-a66e-35f03cc1a1bd/operator/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.493504 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2s825_e72bf30c-6c74-46a6-b215-ac5b1f33c144/kube-rbac-proxy/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.513856 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d77b94747-2s825_e72bf30c-6c74-46a6-b215-ac5b1f33c144/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.632267 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-646778448f-7d5mc_cf2135a7-5206-498c-9391-379fae5f4bfa/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.656091 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fkb27_04b644f3-2f69-4cb5-8ade-04f4be267255/kube-rbac-proxy/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.794168 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-fkb27_04b644f3-2f69-4cb5-8ade-04f4be267255/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.822159 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-8sjts_523ab37c-4bb4-4d00-8134-ad63c2833907/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.826210 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd6c7f4c8-8sjts_523ab37c-4bb4-4d00-8134-ad63c2833907/kube-rbac-proxy/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.985652 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-65qdq_0fa640fb-a493-4f84-bfcf-a3824ddc061f/manager/0.log" Nov 26 07:30:51 crc kubenswrapper[4775]: I1126 07:30:51.992881 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-656dcb59d4-65qdq_0fa640fb-a493-4f84-bfcf-a3824ddc061f/kube-rbac-proxy/0.log" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.314252 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kd5xl"] Nov 26 07:30:52 crc kubenswrapper[4775]: E1126 07:30:52.314913 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13da712-45ab-43fa-9435-38a7f111f60a" containerName="collect-profiles" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.314930 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13da712-45ab-43fa-9435-38a7f111f60a" containerName="collect-profiles" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.315135 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13da712-45ab-43fa-9435-38a7f111f60a" containerName="collect-profiles" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.316526 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.328486 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kd5xl"] Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.427495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54b4t\" (UniqueName: \"kubernetes.io/projected/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-kube-api-access-54b4t\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.427561 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-utilities\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.427695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-catalog-content\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.529691 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54b4t\" (UniqueName: \"kubernetes.io/projected/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-kube-api-access-54b4t\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.529774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-utilities\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.529856 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-catalog-content\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.530309 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-utilities\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.530373 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-catalog-content\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.550475 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54b4t\" (UniqueName: \"kubernetes.io/projected/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-kube-api-access-54b4t\") pod \"certified-operators-kd5xl\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:52 crc kubenswrapper[4775]: I1126 07:30:52.634573 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:30:53 crc kubenswrapper[4775]: I1126 07:30:53.200435 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kd5xl"] Nov 26 07:30:53 crc kubenswrapper[4775]: W1126 07:30:53.207144 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4835093b_16e1_4df3_972f_2eb3ffbe8c2f.slice/crio-8b307dcfe2dcb41e392fa2305b0856eed42b5995fa59aa9b0c4807b5ddad7066 WatchSource:0}: Error finding container 8b307dcfe2dcb41e392fa2305b0856eed42b5995fa59aa9b0c4807b5ddad7066: Status 404 returned error can't find the container with id 8b307dcfe2dcb41e392fa2305b0856eed42b5995fa59aa9b0c4807b5ddad7066 Nov 26 07:30:54 crc kubenswrapper[4775]: I1126 07:30:54.043326 4775 generic.go:334] "Generic (PLEG): container finished" podID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerID="19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689" exitCode=0 Nov 26 07:30:54 crc kubenswrapper[4775]: I1126 07:30:54.043386 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd5xl" event={"ID":"4835093b-16e1-4df3-972f-2eb3ffbe8c2f","Type":"ContainerDied","Data":"19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689"} Nov 26 07:30:54 crc kubenswrapper[4775]: I1126 07:30:54.043657 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd5xl" event={"ID":"4835093b-16e1-4df3-972f-2eb3ffbe8c2f","Type":"ContainerStarted","Data":"8b307dcfe2dcb41e392fa2305b0856eed42b5995fa59aa9b0c4807b5ddad7066"} Nov 26 07:30:54 crc kubenswrapper[4775]: I1126 07:30:54.045769 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 07:30:56 crc kubenswrapper[4775]: I1126 07:30:56.062663 4775 generic.go:334] "Generic (PLEG): container finished" podID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerID="0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b" exitCode=0 Nov 26 07:30:56 crc kubenswrapper[4775]: I1126 07:30:56.062752 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd5xl" event={"ID":"4835093b-16e1-4df3-972f-2eb3ffbe8c2f","Type":"ContainerDied","Data":"0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b"} Nov 26 07:30:58 crc kubenswrapper[4775]: I1126 07:30:58.116403 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd5xl" event={"ID":"4835093b-16e1-4df3-972f-2eb3ffbe8c2f","Type":"ContainerStarted","Data":"7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087"} Nov 26 07:30:58 crc kubenswrapper[4775]: I1126 07:30:58.136122 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kd5xl" podStartSLOduration=3.645975521 podStartE2EDuration="6.13610438s" podCreationTimestamp="2025-11-26 07:30:52 +0000 UTC" firstStartedPulling="2025-11-26 07:30:54.045485888 +0000 UTC m=+3937.406789840" lastFinishedPulling="2025-11-26 07:30:56.535614747 +0000 UTC m=+3939.896918699" observedRunningTime="2025-11-26 07:30:58.1331401 +0000 UTC m=+3941.494444042" watchObservedRunningTime="2025-11-26 07:30:58.13610438 +0000 UTC m=+3941.497408332" Nov 26 07:31:02 crc kubenswrapper[4775]: I1126 07:31:02.634896 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:31:02 crc kubenswrapper[4775]: I1126 07:31:02.635343 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:31:02 crc kubenswrapper[4775]: I1126 07:31:02.686153 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:31:03 crc kubenswrapper[4775]: I1126 07:31:03.207230 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:31:03 crc kubenswrapper[4775]: I1126 07:31:03.266700 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kd5xl"] Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.172182 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kd5xl" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="registry-server" containerID="cri-o://7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087" gracePeriod=2 Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.678786 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.793975 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-utilities\") pod \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.794174 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-catalog-content\") pod \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.794227 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54b4t\" (UniqueName: \"kubernetes.io/projected/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-kube-api-access-54b4t\") pod \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\" (UID: \"4835093b-16e1-4df3-972f-2eb3ffbe8c2f\") " Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.795574 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-utilities" (OuterVolumeSpecName: "utilities") pod "4835093b-16e1-4df3-972f-2eb3ffbe8c2f" (UID: "4835093b-16e1-4df3-972f-2eb3ffbe8c2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.807935 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-kube-api-access-54b4t" (OuterVolumeSpecName: "kube-api-access-54b4t") pod "4835093b-16e1-4df3-972f-2eb3ffbe8c2f" (UID: "4835093b-16e1-4df3-972f-2eb3ffbe8c2f"). InnerVolumeSpecName "kube-api-access-54b4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.848785 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4835093b-16e1-4df3-972f-2eb3ffbe8c2f" (UID: "4835093b-16e1-4df3-972f-2eb3ffbe8c2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.896454 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.896509 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54b4t\" (UniqueName: \"kubernetes.io/projected/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-kube-api-access-54b4t\") on node \"crc\" DevicePath \"\"" Nov 26 07:31:05 crc kubenswrapper[4775]: I1126 07:31:05.896526 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4835093b-16e1-4df3-972f-2eb3ffbe8c2f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.184811 4775 generic.go:334] "Generic (PLEG): container finished" podID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerID="7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087" exitCode=0 Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.184863 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd5xl" event={"ID":"4835093b-16e1-4df3-972f-2eb3ffbe8c2f","Type":"ContainerDied","Data":"7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087"} Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.184928 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kd5xl" event={"ID":"4835093b-16e1-4df3-972f-2eb3ffbe8c2f","Type":"ContainerDied","Data":"8b307dcfe2dcb41e392fa2305b0856eed42b5995fa59aa9b0c4807b5ddad7066"} Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.184958 4775 scope.go:117] "RemoveContainer" containerID="7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.184972 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kd5xl" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.207010 4775 scope.go:117] "RemoveContainer" containerID="0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.228765 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kd5xl"] Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.237990 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kd5xl"] Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.581351 4775 scope.go:117] "RemoveContainer" containerID="19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.718222 4775 scope.go:117] "RemoveContainer" containerID="7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087" Nov 26 07:31:06 crc kubenswrapper[4775]: E1126 07:31:06.718594 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087\": container with ID starting with 7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087 not found: ID does not exist" containerID="7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.718731 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087"} err="failed to get container status \"7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087\": rpc error: code = NotFound desc = could not find container \"7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087\": container with ID starting with 7090ad6ac58d972075152f3c11eb43485537d015f412995a15b3a01507090087 not found: ID does not exist" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.718884 4775 scope.go:117] "RemoveContainer" containerID="0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b" Nov 26 07:31:06 crc kubenswrapper[4775]: E1126 07:31:06.719211 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b\": container with ID starting with 0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b not found: ID does not exist" containerID="0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.719240 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b"} err="failed to get container status \"0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b\": rpc error: code = NotFound desc = could not find container \"0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b\": container with ID starting with 0b9d8a42fee78e18c673d191ad5e10c7e6cd8e8e723b8d355c224653b97c761b not found: ID does not exist" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.719259 4775 scope.go:117] "RemoveContainer" containerID="19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689" Nov 26 07:31:06 crc kubenswrapper[4775]: E1126 07:31:06.719518 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689\": container with ID starting with 19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689 not found: ID does not exist" containerID="19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689" Nov 26 07:31:06 crc kubenswrapper[4775]: I1126 07:31:06.719638 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689"} err="failed to get container status \"19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689\": rpc error: code = NotFound desc = could not find container \"19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689\": container with ID starting with 19508ac34ff513e82419958f695e2bcf206fb96125a2bdbc8affa3a80dc6d689 not found: ID does not exist" Nov 26 07:31:07 crc kubenswrapper[4775]: I1126 07:31:07.339754 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" path="/var/lib/kubelet/pods/4835093b-16e1-4df3-972f-2eb3ffbe8c2f/volumes" Nov 26 07:31:12 crc kubenswrapper[4775]: I1126 07:31:12.212761 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-x5x92_cb8fd937-d759-4e21-98af-5b833a8b3f52/control-plane-machine-set-operator/0.log" Nov 26 07:31:12 crc kubenswrapper[4775]: I1126 07:31:12.377525 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p4qsm_74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb/kube-rbac-proxy/0.log" Nov 26 07:31:12 crc kubenswrapper[4775]: I1126 07:31:12.421180 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-p4qsm_74ed8b1c-7a6e-4cf1-9289-6fffa7932cdb/machine-api-operator/0.log" Nov 26 07:31:24 crc kubenswrapper[4775]: I1126 07:31:24.556830 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-f9chx_4fe2d6c0-b093-452f-8dfa-0db642716f10/cert-manager-controller/0.log" Nov 26 07:31:24 crc kubenswrapper[4775]: I1126 07:31:24.713916 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-99ngt_1d540303-134e-4ca8-9597-9eb8dda29ef4/cert-manager-cainjector/0.log" Nov 26 07:31:24 crc kubenswrapper[4775]: I1126 07:31:24.792068 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-mn5gf_8b00ec65-b72b-4b6c-ac2e-f10f9ddae6cd/cert-manager-webhook/0.log" Nov 26 07:31:36 crc kubenswrapper[4775]: I1126 07:31:36.079801 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-k5dwr_d1e0c343-feef-4493-bec5-d725f665b9d6/nmstate-console-plugin/0.log" Nov 26 07:31:36 crc kubenswrapper[4775]: I1126 07:31:36.282846 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2tzgw_98657427-74ec-48e6-89d9-ef27646c475a/nmstate-handler/0.log" Nov 26 07:31:36 crc kubenswrapper[4775]: I1126 07:31:36.319684 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-2xn6f_910b3c41-e6e8-4c07-945d-d9c580ccce6c/kube-rbac-proxy/0.log" Nov 26 07:31:36 crc kubenswrapper[4775]: I1126 07:31:36.341290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-2xn6f_910b3c41-e6e8-4c07-945d-d9c580ccce6c/nmstate-metrics/0.log" Nov 26 07:31:36 crc kubenswrapper[4775]: I1126 07:31:36.505349 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-jhrbn_5b010a92-1985-467a-9d70-35d65f48f13d/nmstate-operator/0.log" Nov 26 07:31:36 crc kubenswrapper[4775]: I1126 07:31:36.558241 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-c49w5_2f477e8a-59b8-460b-b84d-6b1a4fd50827/nmstate-webhook/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.370039 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-lnhds_2f366115-1210-4d05-9f99-b00eede5cee4/kube-rbac-proxy/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.498759 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-lnhds_2f366115-1210-4d05-9f99-b00eede5cee4/controller/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.610328 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.771183 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.814575 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.819551 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.826314 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:31:50 crc kubenswrapper[4775]: I1126 07:31:50.987972 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.005684 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.008049 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.008212 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.227397 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-metrics/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.239054 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-frr-files/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.242751 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/cp-reloader/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.252500 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/controller/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.446004 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/kube-rbac-proxy/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.457446 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/frr-metrics/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.499243 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/kube-rbac-proxy-frr/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.733130 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/reloader/0.log" Nov 26 07:31:51 crc kubenswrapper[4775]: I1126 07:31:51.747044 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-r6bx2_dccc8029-10ed-427e-a26c-74427323350b/frr-k8s-webhook-server/0.log" Nov 26 07:31:52 crc kubenswrapper[4775]: I1126 07:31:52.202268 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79f98797bf-xz98r_8e93463e-dc10-48b2-b359-494a53052995/manager/0.log" Nov 26 07:31:52 crc kubenswrapper[4775]: I1126 07:31:52.367613 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-766d5f58fc-d4gll_e0758bd1-82f1-4a2f-a009-75975267444c/webhook-server/0.log" Nov 26 07:31:52 crc kubenswrapper[4775]: I1126 07:31:52.458393 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42hxm_953ed086-d444-4c86-9217-134ca8126184/kube-rbac-proxy/0.log" Nov 26 07:31:53 crc kubenswrapper[4775]: I1126 07:31:53.190396 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-r5p7b_d6d95a8e-b6ca-49a7-9344-872cddc08f75/frr/0.log" Nov 26 07:31:53 crc kubenswrapper[4775]: I1126 07:31:53.309129 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-42hxm_953ed086-d444-4c86-9217-134ca8126184/speaker/0.log" Nov 26 07:32:05 crc kubenswrapper[4775]: I1126 07:32:05.885925 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/util/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.055695 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/util/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.116350 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/pull/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.116890 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/pull/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.329097 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/extract/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.335885 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/pull/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.349854 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772e6gs57_4ba6f052-9efe-4347-9394-922477db3e0e/util/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.515015 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-utilities/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.685739 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-content/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.711311 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-utilities/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.757492 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-content/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.871412 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-content/0.log" Nov 26 07:32:06 crc kubenswrapper[4775]: I1126 07:32:06.894065 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/extract-utilities/0.log" Nov 26 07:32:07 crc kubenswrapper[4775]: I1126 07:32:07.077583 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-utilities/0.log" Nov 26 07:32:07 crc kubenswrapper[4775]: I1126 07:32:07.230111 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-utilities/0.log" Nov 26 07:32:07 crc kubenswrapper[4775]: I1126 07:32:07.488127 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-content/0.log" Nov 26 07:32:07 crc kubenswrapper[4775]: I1126 07:32:07.627640 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-content/0.log" Nov 26 07:32:07 crc kubenswrapper[4775]: I1126 07:32:07.893555 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-utilities/0.log" Nov 26 07:32:07 crc kubenswrapper[4775]: I1126 07:32:07.971349 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/extract-content/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.047404 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-td9vf_7dd691a9-d0e3-4241-892d-cf2ff35959c1/registry-server/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.176907 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/util/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.366331 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/pull/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.390310 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/util/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.461037 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/pull/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.512379 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-mhsdl_267d77e2-35df-457c-a0f8-9e7120db903d/registry-server/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.585786 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/util/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.814737 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/pull/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.826805 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68p84z_b4e29c14-21e8-4dc0-a701-493ae7fbda58/extract/0.log" Nov 26 07:32:08 crc kubenswrapper[4775]: I1126 07:32:08.888344 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sh8wn_a4c5e343-455f-43ad-9936-8b64224702bf/marketplace-operator/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.066727 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-utilities/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.236864 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-content/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.252859 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-content/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.293149 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-utilities/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.477967 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-content/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.486557 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/extract-utilities/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.661184 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-2ntql_2c8e888d-5042-4841-8b5e-18efc0d09506/registry-server/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.724515 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-utilities/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.885562 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-utilities/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.891325 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-content/0.log" Nov 26 07:32:09 crc kubenswrapper[4775]: I1126 07:32:09.910182 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-content/0.log" Nov 26 07:32:10 crc kubenswrapper[4775]: I1126 07:32:10.032574 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-content/0.log" Nov 26 07:32:10 crc kubenswrapper[4775]: I1126 07:32:10.047421 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/extract-utilities/0.log" Nov 26 07:32:10 crc kubenswrapper[4775]: I1126 07:32:10.603573 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n8bhs_f1786ea5-82b6-4354-98f5-ae908718294a/registry-server/0.log" Nov 26 07:32:21 crc kubenswrapper[4775]: I1126 07:32:21.420148 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:32:21 crc kubenswrapper[4775]: I1126 07:32:21.420744 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:32:51 crc kubenswrapper[4775]: I1126 07:32:51.420500 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:32:51 crc kubenswrapper[4775]: I1126 07:32:51.422784 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:33:21 crc kubenswrapper[4775]: I1126 07:33:21.419689 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:33:21 crc kubenswrapper[4775]: I1126 07:33:21.420117 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:33:21 crc kubenswrapper[4775]: I1126 07:33:21.420156 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:33:21 crc kubenswrapper[4775]: I1126 07:33:21.420886 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"988d3d41547c245627cb395ac65ae477aa64ca8ebb71a5826ab044045474b18e"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:33:21 crc kubenswrapper[4775]: I1126 07:33:21.420938 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://988d3d41547c245627cb395ac65ae477aa64ca8ebb71a5826ab044045474b18e" gracePeriod=600 Nov 26 07:33:22 crc kubenswrapper[4775]: I1126 07:33:22.432273 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="988d3d41547c245627cb395ac65ae477aa64ca8ebb71a5826ab044045474b18e" exitCode=0 Nov 26 07:33:22 crc kubenswrapper[4775]: I1126 07:33:22.433039 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"988d3d41547c245627cb395ac65ae477aa64ca8ebb71a5826ab044045474b18e"} Nov 26 07:33:22 crc kubenswrapper[4775]: I1126 07:33:22.433159 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerStarted","Data":"c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8"} Nov 26 07:33:22 crc kubenswrapper[4775]: I1126 07:33:22.433196 4775 scope.go:117] "RemoveContainer" containerID="615e4be7b2e02217d29e0fe2ba62c3a8939dd56e547bd444225134c15f948531" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.787706 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cxpgq"] Nov 26 07:33:27 crc kubenswrapper[4775]: E1126 07:33:27.788729 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="extract-utilities" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.788746 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="extract-utilities" Nov 26 07:33:27 crc kubenswrapper[4775]: E1126 07:33:27.788765 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="registry-server" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.788774 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="registry-server" Nov 26 07:33:27 crc kubenswrapper[4775]: E1126 07:33:27.788813 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="extract-content" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.788822 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="extract-content" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.789083 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4835093b-16e1-4df3-972f-2eb3ffbe8c2f" containerName="registry-server" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.795008 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.826674 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cxpgq"] Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.853776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-catalog-content\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.854045 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-utilities\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.854523 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5bbt\" (UniqueName: \"kubernetes.io/projected/8daf0106-59df-49e3-b501-09657a6730aa-kube-api-access-f5bbt\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.956550 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5bbt\" (UniqueName: \"kubernetes.io/projected/8daf0106-59df-49e3-b501-09657a6730aa-kube-api-access-f5bbt\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.956847 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-catalog-content\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.956902 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-utilities\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.957310 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-catalog-content\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.957344 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-utilities\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:27 crc kubenswrapper[4775]: I1126 07:33:27.980855 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5bbt\" (UniqueName: \"kubernetes.io/projected/8daf0106-59df-49e3-b501-09657a6730aa-kube-api-access-f5bbt\") pod \"redhat-operators-cxpgq\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:28 crc kubenswrapper[4775]: I1126 07:33:28.128355 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:28 crc kubenswrapper[4775]: I1126 07:33:28.600609 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cxpgq"] Nov 26 07:33:29 crc kubenswrapper[4775]: I1126 07:33:29.502391 4775 generic.go:334] "Generic (PLEG): container finished" podID="8daf0106-59df-49e3-b501-09657a6730aa" containerID="e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3" exitCode=0 Nov 26 07:33:29 crc kubenswrapper[4775]: I1126 07:33:29.502743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerDied","Data":"e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3"} Nov 26 07:33:29 crc kubenswrapper[4775]: I1126 07:33:29.502786 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerStarted","Data":"efc9544719315fa8aa2fe0a4eb3147e0d0100dd45248213a513f520ec633cb4d"} Nov 26 07:33:30 crc kubenswrapper[4775]: I1126 07:33:30.513735 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerStarted","Data":"aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6"} Nov 26 07:33:31 crc kubenswrapper[4775]: I1126 07:33:31.528191 4775 generic.go:334] "Generic (PLEG): container finished" podID="8daf0106-59df-49e3-b501-09657a6730aa" containerID="aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6" exitCode=0 Nov 26 07:33:31 crc kubenswrapper[4775]: I1126 07:33:31.529121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerDied","Data":"aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6"} Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.543758 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerStarted","Data":"601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c"} Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.569101 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mjlf2"] Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.572053 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.579146 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjlf2"] Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.579896 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cxpgq" podStartSLOduration=3.143175495 podStartE2EDuration="5.579883432s" podCreationTimestamp="2025-11-26 07:33:27 +0000 UTC" firstStartedPulling="2025-11-26 07:33:29.505181115 +0000 UTC m=+4092.866485107" lastFinishedPulling="2025-11-26 07:33:31.941889052 +0000 UTC m=+4095.303193044" observedRunningTime="2025-11-26 07:33:32.579174123 +0000 UTC m=+4095.940478095" watchObservedRunningTime="2025-11-26 07:33:32.579883432 +0000 UTC m=+4095.941187404" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.638589 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-utilities\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.638656 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82pnh\" (UniqueName: \"kubernetes.io/projected/491fc59c-5b67-4797-847b-537602dd9c92-kube-api-access-82pnh\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.638754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-catalog-content\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.747155 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-catalog-content\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.747501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-utilities\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.747602 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82pnh\" (UniqueName: \"kubernetes.io/projected/491fc59c-5b67-4797-847b-537602dd9c92-kube-api-access-82pnh\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.748801 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-utilities\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.749246 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-catalog-content\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.779755 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82pnh\" (UniqueName: \"kubernetes.io/projected/491fc59c-5b67-4797-847b-537602dd9c92-kube-api-access-82pnh\") pod \"redhat-marketplace-mjlf2\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:32 crc kubenswrapper[4775]: I1126 07:33:32.897137 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:33 crc kubenswrapper[4775]: I1126 07:33:33.381486 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjlf2"] Nov 26 07:33:33 crc kubenswrapper[4775]: I1126 07:33:33.561094 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjlf2" event={"ID":"491fc59c-5b67-4797-847b-537602dd9c92","Type":"ContainerStarted","Data":"3574f4d4191d3c3916986d1becac7906681fd68fd98e1132bbb141d614bbab37"} Nov 26 07:33:34 crc kubenswrapper[4775]: I1126 07:33:34.575374 4775 generic.go:334] "Generic (PLEG): container finished" podID="491fc59c-5b67-4797-847b-537602dd9c92" containerID="ffe798c97c3358459640482bff3dc8ba8914cb7037a52bd534b92cbc7484bc12" exitCode=0 Nov 26 07:33:34 crc kubenswrapper[4775]: I1126 07:33:34.575472 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjlf2" event={"ID":"491fc59c-5b67-4797-847b-537602dd9c92","Type":"ContainerDied","Data":"ffe798c97c3358459640482bff3dc8ba8914cb7037a52bd534b92cbc7484bc12"} Nov 26 07:33:36 crc kubenswrapper[4775]: I1126 07:33:36.593695 4775 generic.go:334] "Generic (PLEG): container finished" podID="491fc59c-5b67-4797-847b-537602dd9c92" containerID="8308216d70565fb4a452ce41110f40894bf70627074e74e8af0bfe0bd8df07ce" exitCode=0 Nov 26 07:33:36 crc kubenswrapper[4775]: I1126 07:33:36.593773 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjlf2" event={"ID":"491fc59c-5b67-4797-847b-537602dd9c92","Type":"ContainerDied","Data":"8308216d70565fb4a452ce41110f40894bf70627074e74e8af0bfe0bd8df07ce"} Nov 26 07:33:38 crc kubenswrapper[4775]: I1126 07:33:38.128868 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:38 crc kubenswrapper[4775]: I1126 07:33:38.130878 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:38 crc kubenswrapper[4775]: I1126 07:33:38.187637 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:38 crc kubenswrapper[4775]: I1126 07:33:38.611456 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjlf2" event={"ID":"491fc59c-5b67-4797-847b-537602dd9c92","Type":"ContainerStarted","Data":"4070cb8fda410cce59b7650096fbef5af3b8e8a9c402ddfbad9af1ca17944e57"} Nov 26 07:33:38 crc kubenswrapper[4775]: I1126 07:33:38.631045 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mjlf2" podStartSLOduration=4.113157904 podStartE2EDuration="6.631026402s" podCreationTimestamp="2025-11-26 07:33:32 +0000 UTC" firstStartedPulling="2025-11-26 07:33:34.578050966 +0000 UTC m=+4097.939354938" lastFinishedPulling="2025-11-26 07:33:37.095919474 +0000 UTC m=+4100.457223436" observedRunningTime="2025-11-26 07:33:38.625648207 +0000 UTC m=+4101.986952179" watchObservedRunningTime="2025-11-26 07:33:38.631026402 +0000 UTC m=+4101.992330354" Nov 26 07:33:38 crc kubenswrapper[4775]: I1126 07:33:38.660901 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:39 crc kubenswrapper[4775]: I1126 07:33:39.953934 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cxpgq"] Nov 26 07:33:41 crc kubenswrapper[4775]: I1126 07:33:41.641473 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cxpgq" podUID="8daf0106-59df-49e3-b501-09657a6730aa" containerName="registry-server" containerID="cri-o://601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c" gracePeriod=2 Nov 26 07:33:41 crc kubenswrapper[4775]: E1126 07:33:41.848708 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8daf0106_59df_49e3_b501_09657a6730aa.slice/crio-601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c.scope\": RecentStats: unable to find data in memory cache]" Nov 26 07:33:42 crc kubenswrapper[4775]: I1126 07:33:42.897968 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:42 crc kubenswrapper[4775]: I1126 07:33:42.898268 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:42 crc kubenswrapper[4775]: I1126 07:33:42.960844 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.264142 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.405415 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-catalog-content\") pod \"8daf0106-59df-49e3-b501-09657a6730aa\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.405481 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5bbt\" (UniqueName: \"kubernetes.io/projected/8daf0106-59df-49e3-b501-09657a6730aa-kube-api-access-f5bbt\") pod \"8daf0106-59df-49e3-b501-09657a6730aa\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.405543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-utilities\") pod \"8daf0106-59df-49e3-b501-09657a6730aa\" (UID: \"8daf0106-59df-49e3-b501-09657a6730aa\") " Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.406812 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-utilities" (OuterVolumeSpecName: "utilities") pod "8daf0106-59df-49e3-b501-09657a6730aa" (UID: "8daf0106-59df-49e3-b501-09657a6730aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.411862 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8daf0106-59df-49e3-b501-09657a6730aa-kube-api-access-f5bbt" (OuterVolumeSpecName: "kube-api-access-f5bbt") pod "8daf0106-59df-49e3-b501-09657a6730aa" (UID: "8daf0106-59df-49e3-b501-09657a6730aa"). InnerVolumeSpecName "kube-api-access-f5bbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.508352 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5bbt\" (UniqueName: \"kubernetes.io/projected/8daf0106-59df-49e3-b501-09657a6730aa-kube-api-access-f5bbt\") on node \"crc\" DevicePath \"\"" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.508606 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.522167 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8daf0106-59df-49e3-b501-09657a6730aa" (UID: "8daf0106-59df-49e3-b501-09657a6730aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.610698 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8daf0106-59df-49e3-b501-09657a6730aa-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.660611 4775 generic.go:334] "Generic (PLEG): container finished" podID="8daf0106-59df-49e3-b501-09657a6730aa" containerID="601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c" exitCode=0 Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.660677 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cxpgq" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.660763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerDied","Data":"601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c"} Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.660794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cxpgq" event={"ID":"8daf0106-59df-49e3-b501-09657a6730aa","Type":"ContainerDied","Data":"efc9544719315fa8aa2fe0a4eb3147e0d0100dd45248213a513f520ec633cb4d"} Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.660814 4775 scope.go:117] "RemoveContainer" containerID="601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.690280 4775 scope.go:117] "RemoveContainer" containerID="aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.709655 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cxpgq"] Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.716666 4775 scope.go:117] "RemoveContainer" containerID="e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.720334 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.722368 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cxpgq"] Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.768510 4775 scope.go:117] "RemoveContainer" containerID="601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c" Nov 26 07:33:43 crc kubenswrapper[4775]: E1126 07:33:43.769170 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c\": container with ID starting with 601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c not found: ID does not exist" containerID="601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.769212 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c"} err="failed to get container status \"601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c\": rpc error: code = NotFound desc = could not find container \"601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c\": container with ID starting with 601372269ea92eeda89ade13f94da889f51cebf46e6da131851fbba6c9c0712c not found: ID does not exist" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.769245 4775 scope.go:117] "RemoveContainer" containerID="aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6" Nov 26 07:33:43 crc kubenswrapper[4775]: E1126 07:33:43.769668 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6\": container with ID starting with aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6 not found: ID does not exist" containerID="aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.769703 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6"} err="failed to get container status \"aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6\": rpc error: code = NotFound desc = could not find container \"aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6\": container with ID starting with aea410ee926b6301fc02e9f0d194bd322b4533f8f584ef1bfe5e8d6af22533f6 not found: ID does not exist" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.769811 4775 scope.go:117] "RemoveContainer" containerID="e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3" Nov 26 07:33:43 crc kubenswrapper[4775]: E1126 07:33:43.770063 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3\": container with ID starting with e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3 not found: ID does not exist" containerID="e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3" Nov 26 07:33:43 crc kubenswrapper[4775]: I1126 07:33:43.770084 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3"} err="failed to get container status \"e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3\": rpc error: code = NotFound desc = could not find container \"e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3\": container with ID starting with e3144000c3ac7b360edbd1a0f549c7495c2273a7bb5dee4c7e251a11a71c76d3 not found: ID does not exist" Nov 26 07:33:45 crc kubenswrapper[4775]: I1126 07:33:45.341091 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8daf0106-59df-49e3-b501-09657a6730aa" path="/var/lib/kubelet/pods/8daf0106-59df-49e3-b501-09657a6730aa/volumes" Nov 26 07:33:46 crc kubenswrapper[4775]: I1126 07:33:46.150583 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjlf2"] Nov 26 07:33:46 crc kubenswrapper[4775]: I1126 07:33:46.150803 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mjlf2" podUID="491fc59c-5b67-4797-847b-537602dd9c92" containerName="registry-server" containerID="cri-o://4070cb8fda410cce59b7650096fbef5af3b8e8a9c402ddfbad9af1ca17944e57" gracePeriod=2 Nov 26 07:33:46 crc kubenswrapper[4775]: I1126 07:33:46.692634 4775 generic.go:334] "Generic (PLEG): container finished" podID="491fc59c-5b67-4797-847b-537602dd9c92" containerID="4070cb8fda410cce59b7650096fbef5af3b8e8a9c402ddfbad9af1ca17944e57" exitCode=0 Nov 26 07:33:46 crc kubenswrapper[4775]: I1126 07:33:46.692672 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjlf2" event={"ID":"491fc59c-5b67-4797-847b-537602dd9c92","Type":"ContainerDied","Data":"4070cb8fda410cce59b7650096fbef5af3b8e8a9c402ddfbad9af1ca17944e57"} Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.123998 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.183618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-catalog-content\") pod \"491fc59c-5b67-4797-847b-537602dd9c92\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.184036 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-utilities\") pod \"491fc59c-5b67-4797-847b-537602dd9c92\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.184187 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82pnh\" (UniqueName: \"kubernetes.io/projected/491fc59c-5b67-4797-847b-537602dd9c92-kube-api-access-82pnh\") pod \"491fc59c-5b67-4797-847b-537602dd9c92\" (UID: \"491fc59c-5b67-4797-847b-537602dd9c92\") " Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.185820 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-utilities" (OuterVolumeSpecName: "utilities") pod "491fc59c-5b67-4797-847b-537602dd9c92" (UID: "491fc59c-5b67-4797-847b-537602dd9c92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.192283 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/491fc59c-5b67-4797-847b-537602dd9c92-kube-api-access-82pnh" (OuterVolumeSpecName: "kube-api-access-82pnh") pod "491fc59c-5b67-4797-847b-537602dd9c92" (UID: "491fc59c-5b67-4797-847b-537602dd9c92"). InnerVolumeSpecName "kube-api-access-82pnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.219461 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "491fc59c-5b67-4797-847b-537602dd9c92" (UID: "491fc59c-5b67-4797-847b-537602dd9c92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.287982 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.288053 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82pnh\" (UniqueName: \"kubernetes.io/projected/491fc59c-5b67-4797-847b-537602dd9c92-kube-api-access-82pnh\") on node \"crc\" DevicePath \"\"" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.288072 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/491fc59c-5b67-4797-847b-537602dd9c92-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.704423 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mjlf2" event={"ID":"491fc59c-5b67-4797-847b-537602dd9c92","Type":"ContainerDied","Data":"3574f4d4191d3c3916986d1becac7906681fd68fd98e1132bbb141d614bbab37"} Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.704484 4775 scope.go:117] "RemoveContainer" containerID="4070cb8fda410cce59b7650096fbef5af3b8e8a9c402ddfbad9af1ca17944e57" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.704505 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mjlf2" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.737796 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjlf2"] Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.744995 4775 scope.go:117] "RemoveContainer" containerID="8308216d70565fb4a452ce41110f40894bf70627074e74e8af0bfe0bd8df07ce" Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.751007 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mjlf2"] Nov 26 07:33:47 crc kubenswrapper[4775]: I1126 07:33:47.775924 4775 scope.go:117] "RemoveContainer" containerID="ffe798c97c3358459640482bff3dc8ba8914cb7037a52bd534b92cbc7484bc12" Nov 26 07:33:49 crc kubenswrapper[4775]: I1126 07:33:49.344314 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="491fc59c-5b67-4797-847b-537602dd9c92" path="/var/lib/kubelet/pods/491fc59c-5b67-4797-847b-537602dd9c92/volumes" Nov 26 07:33:54 crc kubenswrapper[4775]: I1126 07:33:54.774080 4775 generic.go:334] "Generic (PLEG): container finished" podID="4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6" containerID="880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12" exitCode=0 Nov 26 07:33:54 crc kubenswrapper[4775]: I1126 07:33:54.774907 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wphn7/must-gather-96w2j" event={"ID":"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6","Type":"ContainerDied","Data":"880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12"} Nov 26 07:33:54 crc kubenswrapper[4775]: I1126 07:33:54.775929 4775 scope.go:117] "RemoveContainer" containerID="880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12" Nov 26 07:33:54 crc kubenswrapper[4775]: I1126 07:33:54.856348 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wphn7_must-gather-96w2j_4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6/gather/0.log" Nov 26 07:34:04 crc kubenswrapper[4775]: I1126 07:34:04.196259 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wphn7/must-gather-96w2j"] Nov 26 07:34:04 crc kubenswrapper[4775]: I1126 07:34:04.196984 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wphn7/must-gather-96w2j" podUID="4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6" containerName="copy" containerID="cri-o://8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d" gracePeriod=2 Nov 26 07:34:04 crc kubenswrapper[4775]: I1126 07:34:04.207665 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wphn7/must-gather-96w2j"] Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.166317 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wphn7_must-gather-96w2j_4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6/copy/0.log" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.167271 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.170535 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wphn7_must-gather-96w2j_4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6/copy/0.log" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.171121 4775 generic.go:334] "Generic (PLEG): container finished" podID="4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6" containerID="8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d" exitCode=143 Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.171201 4775 scope.go:117] "RemoveContainer" containerID="8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.171414 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wphn7/must-gather-96w2j" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.222076 4775 scope.go:117] "RemoveContainer" containerID="880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.252694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-must-gather-output\") pod \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.253034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7f4rw\" (UniqueName: \"kubernetes.io/projected/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-kube-api-access-7f4rw\") pod \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\" (UID: \"4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6\") " Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.259017 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-kube-api-access-7f4rw" (OuterVolumeSpecName: "kube-api-access-7f4rw") pod "4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6" (UID: "4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6"). InnerVolumeSpecName "kube-api-access-7f4rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.328803 4775 scope.go:117] "RemoveContainer" containerID="8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d" Nov 26 07:34:05 crc kubenswrapper[4775]: E1126 07:34:05.336850 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d\": container with ID starting with 8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d not found: ID does not exist" containerID="8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.336933 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d"} err="failed to get container status \"8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d\": rpc error: code = NotFound desc = could not find container \"8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d\": container with ID starting with 8b695bd53176e96fab86100ebe3e37f89378be62e250cecff2779bbe74c22b2d not found: ID does not exist" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.336965 4775 scope.go:117] "RemoveContainer" containerID="880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12" Nov 26 07:34:05 crc kubenswrapper[4775]: E1126 07:34:05.341521 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12\": container with ID starting with 880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12 not found: ID does not exist" containerID="880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.341561 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12"} err="failed to get container status \"880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12\": rpc error: code = NotFound desc = could not find container \"880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12\": container with ID starting with 880d329bca1ec03f92479b6d385d9fe089d92822ad5a711af6a777a5394edc12 not found: ID does not exist" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.358280 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7f4rw\" (UniqueName: \"kubernetes.io/projected/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-kube-api-access-7f4rw\") on node \"crc\" DevicePath \"\"" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.438462 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6" (UID: "4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 07:34:05 crc kubenswrapper[4775]: I1126 07:34:05.460027 4775 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 26 07:34:07 crc kubenswrapper[4775]: I1126 07:34:07.353241 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6" path="/var/lib/kubelet/pods/4f7a21c1-1616-4ea8-8f7f-5712d7b3e4c6/volumes" Nov 26 07:35:21 crc kubenswrapper[4775]: I1126 07:35:21.419574 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:35:21 crc kubenswrapper[4775]: I1126 07:35:21.420187 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:35:51 crc kubenswrapper[4775]: I1126 07:35:51.420555 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:35:51 crc kubenswrapper[4775]: I1126 07:35:51.421296 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.419517 4775 patch_prober.go:28] interesting pod/machine-config-daemon-p7z2d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.421277 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.421506 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.422900 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8"} pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.423112 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerName="machine-config-daemon" containerID="cri-o://c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" gracePeriod=600 Nov 26 07:36:21 crc kubenswrapper[4775]: E1126 07:36:21.563057 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.604640 4775 generic.go:334] "Generic (PLEG): container finished" podID="687d9603-b248-4ad7-95d6-a23c110afaf1" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" exitCode=0 Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.604682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" event={"ID":"687d9603-b248-4ad7-95d6-a23c110afaf1","Type":"ContainerDied","Data":"c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8"} Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.604714 4775 scope.go:117] "RemoveContainer" containerID="988d3d41547c245627cb395ac65ae477aa64ca8ebb71a5826ab044045474b18e" Nov 26 07:36:21 crc kubenswrapper[4775]: I1126 07:36:21.605530 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:36:21 crc kubenswrapper[4775]: E1126 07:36:21.607502 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:36:35 crc kubenswrapper[4775]: I1126 07:36:35.332350 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:36:35 crc kubenswrapper[4775]: E1126 07:36:35.333030 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:36:46 crc kubenswrapper[4775]: I1126 07:36:46.328301 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:36:46 crc kubenswrapper[4775]: E1126 07:36:46.329102 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:37:00 crc kubenswrapper[4775]: I1126 07:37:00.328830 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:37:00 crc kubenswrapper[4775]: E1126 07:37:00.329847 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:37:13 crc kubenswrapper[4775]: I1126 07:37:13.328164 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:37:13 crc kubenswrapper[4775]: E1126 07:37:13.330608 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:37:27 crc kubenswrapper[4775]: I1126 07:37:27.333515 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:37:27 crc kubenswrapper[4775]: E1126 07:37:27.334341 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:37:39 crc kubenswrapper[4775]: I1126 07:37:39.327680 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:37:39 crc kubenswrapper[4775]: E1126 07:37:39.328543 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:37:53 crc kubenswrapper[4775]: I1126 07:37:53.328420 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:37:53 crc kubenswrapper[4775]: E1126 07:37:53.329281 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:38:04 crc kubenswrapper[4775]: I1126 07:38:04.328142 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:38:04 crc kubenswrapper[4775]: E1126 07:38:04.329917 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:38:16 crc kubenswrapper[4775]: I1126 07:38:16.328531 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:38:16 crc kubenswrapper[4775]: E1126 07:38:16.329808 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" Nov 26 07:38:27 crc kubenswrapper[4775]: I1126 07:38:27.337254 4775 scope.go:117] "RemoveContainer" containerID="c6ecce7454ebd319e770e56b109a41c7f1491726a16987e1c6e0ec9daab199b8" Nov 26 07:38:27 crc kubenswrapper[4775]: E1126 07:38:27.340154 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-p7z2d_openshift-machine-config-operator(687d9603-b248-4ad7-95d6-a23c110afaf1)\"" pod="openshift-machine-config-operator/machine-config-daemon-p7z2d" podUID="687d9603-b248-4ad7-95d6-a23c110afaf1" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111527204024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111527205017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111516212016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111516212015450 5ustar corecore